penat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f0000000080)='./file0\x00', r0) 00:08:31 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 3: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140)={0xffffffffffffffff}, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) getsockname$netlink(r0, &(0x7f0000000100), &(0x7f0000000140)=0xc) setsockopt$SO_VM_SOCKETS_BUFFER_MAX_SIZE(r0, 0x28, 0x2, &(0x7f00000011c0)=0xbd, 0x8) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:31 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:31 executing program 2: fsopen(&(0x7f0000000000)='proc\x00', 0x0) r0 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f0000000080)='./file0\x00', r0) 00:08:31 executing program 0: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket$kcm(0x29, 0x5, 0x0) r1 = socket$kcm(0x29, 0x7, 0x0) recvmsg$kcm(r1, &(0x7f0000000380)={&(0x7f0000000140)=@in={0x2, 0x0, @remote}, 0x80, &(0x7f0000000300)=[{&(0x7f00000001c0)=""/22, 0x16}, {&(0x7f0000000200)=""/158, 0x9e}, {&(0x7f00000002c0)=""/10, 0xa}], 0x3, &(0x7f0000000340)=""/36, 0x24}, 0xb523f66b0d1d6f64) recvmsg$kcm(r0, &(0x7f0000000640)={0x0, 0x0, 0x0}, 0x0) bind(r0, &(0x7f00000000c0)=@nl=@unspec, 0x80) r2 = bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) bpf$OBJ_PIN_PROG(0x6, &(0x7f0000000840)=@o_path={&(0x7f00000003c0)='./file0\x00', r2}, 0x18) getsockname(r0, &(0x7f0000000000)=@xdp, &(0x7f0000000080)=0x80) 00:08:31 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}], 0x2, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 4: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, 0x0, &(0x7f0000000080)='./file0\x00', r1) 00:08:31 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:31 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:31 executing program 3: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, 0x0, &(0x7f0000000080)='./file0\x00', r1) 00:08:32 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140)={0xffffffffffffffff}, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) getsockname$netlink(r0, &(0x7f0000000100), &(0x7f0000000140)=0xc) setsockopt$SO_VM_SOCKETS_BUFFER_MAX_SIZE(r0, 0x28, 0x2, &(0x7f00000011c0)=0xbd, 0x8) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 2: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, 0x0, &(0x7f0000000080)='./file0\x00', r1) 00:08:32 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:32 executing program 3: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140)={0xffffffffffffffff}, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) getsockname$netlink(r0, &(0x7f0000000100), &(0x7f0000000140)=0xc) r1 = socket(0x10, 0x3, 0x4) sendto$rose(r1, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 4: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', 0x0, r1) 00:08:32 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140)={0xffffffffffffffff}, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) r1 = socket(0x10, 0x3, 0x4) sendto$rose(r1, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 2: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', 0x0, r1) 00:08:32 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140), 0x0) r0 = socket(0x10, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 3: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', 0x0, r1) 00:08:32 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:32 executing program 2: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f0000000080)='./file0\x00', 0xffffffffffffffff) 00:08:32 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) r0 = socket(0x10, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 4: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f0000000080)='./file0\x00', 0xffffffffffffffff) 00:08:32 executing program 1: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket(0x10, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 1: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket$kcm(0x29, 0x5, 0x0) r1 = socket$kcm(0x29, 0x7, 0x0) recvmsg$kcm(r1, &(0x7f0000000380)={&(0x7f0000000140)=@in={0x2, 0x0, @remote}, 0x80, &(0x7f0000000300)=[{&(0x7f00000001c0)=""/22, 0x16}, {&(0x7f0000000200)=""/158, 0x9e}, {&(0x7f00000002c0)=""/10, 0xa}], 0x3, &(0x7f0000000340)=""/36, 0x24}, 0xb523f66b0d1d6f64) recvmsg$kcm(r0, &(0x7f0000000640)={0x0, 0x0, 0x0}, 0x0) bind(r0, &(0x7f00000000c0)=@nl=@unspec, 0x80) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:32 executing program 3: r0 = socket(0x10, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 0: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f0000000080)='./file0\x00', 0xffffffffffffffff) 00:08:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 0: r0 = socket(0x0, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}], 0x2, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:32 executing program 3: r0 = socket(0x0, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x0, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}], 0x2, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket(0x0, 0x3, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = socket(0x10, 0x0, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 0: r0 = socket(0x10, 0x0, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket(0x10, 0x0, 0x4) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 1: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket$kcm(0x29, 0x5, 0x0) r1 = socket$kcm(0x29, 0x7, 0x0) recvmsg$kcm(r1, &(0x7f0000000380)={&(0x7f0000000140)=@in={0x2, 0x0, @remote}, 0x80, &(0x7f0000000300)=[{&(0x7f00000001c0)=""/22, 0x16}, {&(0x7f0000000200)=""/158, 0x9e}, {&(0x7f00000002c0)=""/10, 0xa}], 0x3, &(0x7f0000000340)=""/36, 0x24}, 0xb523f66b0d1d6f64) recvmsg$kcm(r0, &(0x7f0000000640)={0x0, 0x0, 0x0}, 0x0) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:33 executing program 0: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:33 executing program 3: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:33 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:33 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:33 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 0: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) socket$kcm(0x29, 0x5, 0x0) r0 = socket$kcm(0x29, 0x7, 0x0) recvmsg$kcm(r0, &(0x7f0000000380)={&(0x7f0000000140)=@in={0x2, 0x0, @remote}, 0x80, &(0x7f0000000300)=[{&(0x7f00000001c0)=""/22, 0x16}, {&(0x7f0000000200)=""/158, 0x9e}, {&(0x7f00000002c0)=""/10, 0xa}], 0x3, &(0x7f0000000340)=""/36, 0x24}, 0xb523f66b0d1d6f64) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:34 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc", 0x800, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:34 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20}, 0x4000000) 00:08:34 executing program 3: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) socket$kcm(0x29, 0x5, 0x0) socket$kcm(0x29, 0x7, 0x0) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:34 executing program 0: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) socket$kcm(0x29, 0x5, 0x0) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20}, 0x0) 00:08:34 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:34 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:34 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:34 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0xffff, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:34 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:35 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390", 0x400, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:35 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 3: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:35 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 1: r0 = fsopen(&(0x7f0000000000)='proc\x00', 0x0) r1 = openat$nvme_fabrics(0xffffffffffffff9c, &(0x7f00000000c0), 0x0, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', 0x0, r1) 00:08:35 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342e", 0x200, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x118, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:35 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x118, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x11c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x10c, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}]}]}, 0x11c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}], 0x2, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x7fffffff}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x11c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x10c, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}]}]}, 0x11c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x31, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}, @ringbuf_query], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf", 0x100, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x2d, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x2}}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xd8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic]}]}, 0xd8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x24, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}, @call={0x85, 0x0, 0x0, 0x7}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x11c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x10c, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x0, 0x0, 0x0, @u64=0x87}]}]}, 0x11c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xd8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}]}]}, 0xd8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:35 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x23, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, 0x1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x40}}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:35 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:35 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xba, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x1a, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x11c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x10c, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x11c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa", 0x80, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x18, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xb8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xa6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}]}]}, 0xb8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x13, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x12, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xb4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xa2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}]}]}, 0xb4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 1: sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x10, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xac, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x9a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99"]}]}, 0xac}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f6", 0x40, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 0: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e2", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:36 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1}, 0x4000000) 00:08:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e2", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x0, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 1: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934", 0x20, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e2", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1}, 0x0) 00:08:36 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 3: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:36 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:36 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:36 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 1: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040)=@kern={0x10, 0x0, 0x0, 0x40000000}, 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:37 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 3: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x0, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0", 0x10, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:37 executing program 4: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:37 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x118, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x105, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x118}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:37 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 3: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x128, 0x21, 0x1, 0x70bd2c, 0x25dfdbfd, "", [@nested={0x118, 0x123, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21", @typed={0xc, 0x10e, 0x0, 0x0, @u64=0x87}, @typed={0xc, 0x6d, 0x0, 0x0, @u64=0x402}]}]}, 0x128}, {&(0x7f00000001c0)={0x1e0, 0x16, 0x10, 0x70bd27, 0x25dfdbfb, "", [@nested={0x25, 0x23, 0x0, 0x1, [@generic="75fa97a752c795a82c0947911059c698d7b8ecc01cea54d0f61f89a70c5eb5bf1c"]}, @typed={0xd0, 0xe3, 0x0, 0x0, @binary="ca3e1a340808f892e2e0a040e6d95b8994b9d007c7eb6c8bc624a859355fadcf4629ff99bd50063946874cbc927a7d99fdff8ddd2d6543e40317e4286ce2560f6d3ac32cc7c2a01ff06dbd3bef0c3f1a148a8e8f7e3c06b633ad6f8e8ab72abd647a811e2f04c9cd4293169555f7d3a2a669d5fcb1949a89654e927d02d322c1fe7e843d2f9cdfd93dd5cf0a694a814179eee10a449070fa879e1ae428ef4b0a2605690382ce7ce2a1b18cb60b18afea7adc1b9a03d4118d4f83a3a96c6d1c8cf96217ffca35f74579ea2877"}, @typed={0x6, 0x101, 0x0, 0x0, @str='-\x00'}, @nested={0xd0, 0x96, 0x0, 0x1, [@typed={0xbf, 0x83, 0x0, 0x0, @binary="8cd73a834ff5a7c79c2ec72625c957ffe82251940dbe24237d580012eb0266f499f0315c5483c6e0259d689972fa73d4dac7fb83c5a9603c9dcb82655b8a29573fd0c8a5589c2313b43fc201ac5307e64e1bfaaf4054c3c01f0f5c0aa5b3c78c3e47123c944cffff387c5509ad178ca7178a4d0d33f0280ad2d9c964eba03ff932fa06c8051e59995f0356daa79f60467777e56eb5500aac820944db515d681c96660dc0b36442e4f5e2c16e34a1e22065d0bc38d7ef86fb2bbed7"}, @typed={0xc, 0x63, 0x0, 0x0, @u64=0x8000000000000000}]}]}, 0x1e0}, {&(0x7f0000000940)={0x2e4, 0x2f, 0x10, 0x70bd2a, 0x25dfdbfb, "", [@typed={0x4, 0xfc}, @nested={0x24b, 0xd2, 0x0, 0x1, [@generic="fba2f7f29f5ae4158ff90b7613cded456ade8e2403e8d48b26320092305a7474e51014d4596284893a2caa4972528e7cc6235daa1f217c7a02", @generic="dc78da4e67ea9ff4bf6958ec07e3f1a5f2a7d3917ee9fb1ec8aff34676f9c077ccf02f2e1b267d1c6bd8544a5e8118285ee86cb92b8049e0e7dd265a17b915d0211c7e4c0fb326d9361bc7797e6ef8c92707eeec8554077318a39bd14d1d06c624e9e0d0d1bfe5dc76381d545defc2c823e002aa72edfb8bac78a0c504ab37c7c5473adebe8388b259cef4dcf013367345a964ccbe1aa40ca2a528d53df3d26332e877a2e21071ab1dfb7a9c5171a42b2f88a4c39333e9fae357d7ef472c69eac8c89030a9c47ccf9b508ca652adc735fb3086e78de99f4091ae3175a56c4c", @typed={0x15, 0xa3, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="a51a8c1b6a0ce9df9a4a7d42d38e16672e049fe8a71cfc47602b2f742b0039fb15d3bd54398f618e7ad78583106d3ccaa6d92f97a5b1d8b4a477972664259432c112d1a0e0b2fc15ff012382814683e8ef07fd9129e70d211a6a2d1c02de554397474148a40f9b492bca29859ea2eb50ca3af3f1527aeef7a475f81f972d42a42ebd5bda44a599acd7fafec6b5998d8d4a883b8dd4c92dcc5351d03ce818607689c6f38de5a4d2a4ad090e48a0b13245d04f4f17934b59ec1edf3a75d53d2b4eb83c", @typed={0x8, 0x3d, 0x0, 0x0, @ipv4=@local}, @generic="038c65e401ac5cf648e08c6c40a2702f90ef3bf3b50a06dcb1ff30d07ad781e350a97f064d526296b88a622f47fc2323be16c8fe5472c9255b81689099261aaa6f0ff96d15", @typed={0x8, 0x89, 0x0, 0x0, @u32=0x401}]}, @typed={0x15, 0x8f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ba51fb3077e115fc7fe052f66af3a4fc32dc3bcb51d2b33497a8c7f896fee6a076fb882b76e404db1bc49533165632bd31bb7195a83243045bb384b730699047ddf79ecb71da4d99b6a798bb6f1857a50f945dec7701c1c105842c455c47cec2eb76b642670e1daa14a9"]}, 0x2e4}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 1: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, 0x0, 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba7d2d5059", 0x8, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:37 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, 0x0, &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xd8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic]}]}, 0xd8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10, 0x6}, 0x90) 00:08:37 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e073fba", 0x4, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75}, 0x90) [ 509.374729][T26885] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_cmd_wq": -EINTR [ 517.896563][T28080] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:37 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x1a, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [@kfunc={0x85, 0x0, 0x2, 0x0, 0x5}, @map_idx={0x18, 0x3, 0x5, 0x0, 0x10}, @kfunc={0x85, 0x0, 0x2, 0x0, 0x1}, @tail_call={{0x18, 0x2, 0x1, 0x0, 0x1}}, @btf_id={0x18, 0x3, 0x3, 0x0, 0x1}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x41100, 0x36, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, &(0x7f0000000600)={0x8, 0x3}, 0x8, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:37 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:37 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:37 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20}, 0x4000000) 00:08:37 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) [ 518.036812][T28101] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:38 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20}, 0x0) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e07", 0x2, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) [ 518.145728][T28119] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x104, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) [ 518.489472][T28167] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:38 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:38 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) [ 518.599482][T28186] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:38 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:38 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10, 0x6}, 0x90) [ 518.760144][T28207] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:38 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:38 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:38 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:38 executing program 2: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:38 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:38 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:38 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:39 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:39 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:39 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 2: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x0, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) 00:08:39 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xeb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8f", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 519.286462][T28286] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:39 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xeb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8f", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 0: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xeb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8f", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 519.436493][T28312] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.1'. 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 519.597435][T28335] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:39 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) [ 519.634327][T28346] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfd, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c96", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:39 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfd, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c96", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 1: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:39 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfd, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c96", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:40 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af53", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 1: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, 0x0, &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10, 0x6}, 0x90) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af53", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x0, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x4b, &(0x7f0000000680)=""/75}, 0x90) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af53", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x0, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x0, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20}, 0x0) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:40 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf7, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1}, 0x0) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xd8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic]}]}, 0xd8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xd8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}]}]}, 0xd8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xba, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xb8, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xa6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}]}]}, 0xb8}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xb4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xa2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}]}]}, 0xb4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 3: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xac, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x9a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99"]}]}, 0xac}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:41 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:42 executing program 2: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:42 executing program 0: r0 = socket(0x0, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1}, 0x0) 00:08:42 executing program 4: r0 = socket(0x10, 0x0, 0x0) sendto$rose(r0, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:42 executing program 1: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x64, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x53, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624"]}]}, 0x64}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: socket(0x10, 0x3, 0x0) sendto$rose(0xffffffffffffffff, &(0x7f0000000180)="8e", 0x1, 0x0, 0x0, 0x0) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x64, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x53, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624"]}]}, 0x64}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x75, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x64, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x53, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624"]}]}, 0x64}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:42 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x80, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x6d, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x80}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x77, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x80, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x6d, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x80}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:43 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 0: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x0, 0x0, 0x0) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x77, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x80, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x6d, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x80}], 0x1}, 0x4000000) 00:08:43 executing program 1: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x0, 0x0, 0x0) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x77, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:43 executing program 3: r0 = socket(0x10, 0x3, 0x0) sendto$rose(r0, &(0x7f0000000180), 0x0, 0x0, 0x0, 0x0) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x80, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x6d, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x80}], 0x1}, 0x0) 00:08:43 executing program 3: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x9c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x89, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec"]}]}, 0x9c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be865b17b87329d0ec9b21"]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865c"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865c"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865c"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x14, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x4}]}, 0x14}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xef, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) [ 523.875339][T28993] __nla_validate_parse: 12 callbacks suppressed [ 523.875355][T28993] netlink: 120 bytes leftover after parsing attributes in process `syz-executor.1'. 00:08:43 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xba, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x103, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x79, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce2"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:43 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:43 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x79, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce2"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x79, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce2"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7d, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7d, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, 0x0, &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7d, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10, 0x6}, 0x90) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x80, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x6d, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be86", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x80}], 0x1}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x1}, 0x90) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) [ 524.684851][T29101] netlink: 120 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x79, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce2"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 524.822208][T29124] netlink: 120 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) [ 524.905695][T29135] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.0'. [ 524.937839][T29141] netlink: 120 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:44 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:45 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: bpf$PROG_LOAD_XDP(0x5, 0x0, 0x0) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = accept4$vsock_stream(0xffffffffffffffff, &(0x7f0000000000)={0x28, 0x0, 0x2711}, 0x10, 0x800) pipe2(&(0x7f0000000140)={0xffffffffffffffff}, 0x0) sendto$rose(r1, 0x0, 0x0, 0x0, 0x0, 0x0) (async) getsockname$netlink(r1, &(0x7f0000000100), &(0x7f0000000140)=0xc) (async) setsockopt$SO_VM_SOCKETS_BUFFER_MAX_SIZE(r1, 0x28, 0x2, &(0x7f00000011c0)=0xbd, 0x8) (async, rerun: 64) r2 = socket(0x10, 0x3, 0x4) (rerun: 64) sendto$rose(r2, &(0x7f0000000180)="8e073fba7d2d50596e3837c0d933cba0a3079b4f9a2f4851b2793a1375d62934cdb8b9d4e366a58dfe9d29499e9e62dcaf8fd987614ebdb94a0fdfd2769430f61a43788b07805f564a8ac568e0aaec19b5e90936700d124650d294f67cac09320c460b3470c1e3a58e19859995fa33f8bc8542a107afb10f696818c7c612c2aa5f273997a6b527ca1badfe8eb50e112b8293ab4d391d184b3209ec19e8436b125be35f306b970b2b1872b197ce98bcf75335406ed5e8fa1a51eed182abb8edd187c603439ea0f9f651c0f61a6f4349504d8b42f62c99aec08bec202bb3451180cb53168dd6e1ae8a22359c102c0b1e04067464cd781bb167533a3c8414b965cf655da1b9a4c7b5aa8dd37670f569b6a0350fce2e90591ca9aa7ea0836293bbd49f7bbbcef30badf700f88af219bb0178d797f39301b3f1e3896c4aa2bd2b80338fb8da6d4b6ab471fd978ab5a6f8e68f3f93d9e68058c0c4ac2a912faaca8e778f0046c3dad2a5c8eeda6c01928a135e013cc14837bfe489d203024ee5cd2001b86bd169d22533dc051114125de6f4369647bab3373043669f13d7fb4483331944573d536fd2c9afcea8873ed4961876c84e8b01eed1f3235a0a2511219f81377208c77212fda22f68549ec3f0ec7bd2fd12cc9704dc3825a116cc2167211fc30911fbac91f81904c9add28339c0316d2931e64338f128e7950a5896ca28342ee21972804b9b1f91813cb9484eb1d7fa2c53d66fd71c6c6595307e400441e4803f3d5fa1737a9d072b64b5c12a3e35360d47eb7ec745c7e346c97f92f95697917fa5674410aa8236760f0d0ad7d3556d17304a5ef2376a6a73104008efbcb075977f4e5713d1b12ffde4eb6b62766b317e2a50e1f22224584128ede2d7bbbee116bb11a4aedc4de327fc4da1ec47aacf9091f332b78fde7eebe4e38aa68edd0f790ed3c23396868ef1647bb7a6309de9034d8ead6082181982629e03eadec925bcd482fcc2d176813de0a4bb0b55973ebcb4aacf22c40c0c97801b65d5b396dbf937aa724c2ba754d39fa58130794ca256b55e34c9aa444b0db5c6bac7a4d28c51886252db441c415946a2272e871f88632287486845b6708053be65003b835ec02f97750fd112bd268c4bee834e91320c82338c3a213a28f44b8f68063cbf008a23d5c26076d0ea19e3daab9e185b9d05e473df5fc111d571b39321c33ca2bacab8e9ee25a8e1f83ee9803572a55de58465aedfb0307a5cad9d674dfe1e4b0bac39f943ed9f1a2450cb1a079dd8349679094e1258a0c0d74f719f0bcd6d7b39876bff3e9d6c01dc16769a89828d1d1ca253f5531af45b32b98294268234a82dea49106880064a07f4ef63318efe3ccb627e1cbabba75b4985402f89882b574fa98e1e7df555a1121dac93794bcaea2c22b4ebe1a290924668aa9bea05649390fa9408ce07935912f494f39b51df1b9bdbdd8a5023ce68774b7f57c4c6afbe91912b959a450b751ec41f17af517c0e80718dc79b5147f7b3d4f2aeb30fa3119a44072ec484542503d6c6cfc92837aae83271660eb3885d65f57179a4d4d3463db67865bf7b7f5293112bfbe2d39a37cf0ea028c10b910d7729bc2a847656269c22e30621594d38a6c0625e928c3455585398b8c088b58360edcf5e145e356ab43be04d5fbddad737e65d7b76de22da9369d2c9f17608d674ec567e89463303b49cf1aa0925cc235816a4edf79ded6627583a2a038ba45e8c2fe2d13aed4cdf52867322cb97098a8245d2e11bd133fbe7e19cc1d60cd3d48cf70036390728d9643af5761e78291e83abe462421ba24d96de8842685120addc759298dd791b05d86f5ed94c229a50a558ef3a63195eed7b463cfbb75f52170c9fdf84ec62e75707b71cff13c7f89cca537974de2d5c46684fa33cf86fc9a032b99d81976efbc4fb7094c34252d0cab7ac9799a72bf35baa5f3ec0ce0532efd3e6d090fc1e56f9f7ffa51537931e4e330374abaed30889321e5670414294762cf150ca65cf4b35df7c285a1837589f9c44c77bfc2cb9d26f98afc0ad20cfb049eb7a7f3fb281dda44c53c1ef995800145e6d8cf02fa419871a4ad05b399f28d0f67ef2ec15e8f81361517d18a3a5c8af574ae96e1322e2595cd5867a6c58e0daf3e4ddaf884d6c8252e0db5f420db0eb02918ab9c724457cb82f499113d35bd705400ae476c3cf38dfeff1c26e9c27a1cbd9c326dbb5e667568c1895f47de3d6e58779c38f70ae970a65b9f00d006435b83258cfb74d137abca2fda2a3593f0475ef647942c0ed40fc5871438bc8dc94c7ac38d3768996a0d9fbe5e611ba471536045b5f57ea85edd5f5d906d97a50b3f2fd7fe1782cd66a84c740677f8ad22226b128ba52e4a2cd3a4d1bf49568974cf3c119f39c7a336750e3360da194be940ebd6c4fcf73f689e7fd6e3c2c7362c274e75c52d9e951eb457414cc80a7cf29d189f2034756403c6b216b29a97245e93355107e206b404cab1a1206cd2606b67fb65f47a37dd61e3aaa3443554490c7376f5abcd163e1d6551cd2516719829d41811bd1edcad9e16d62386e5d372c717bf337a7d2ddea27720af27c67cd2857839c14b173bb086ddb7958b09e7f0029a9c7a24af68a588f34cb3bc6c4e28d5b36ef6c60f890b14e6f33c176cbc2ddebe8d7317f1c9056cbc83ec0c36338a45127608ab1ee7c6206feb24808806160a004716775bd4566cc2ad1ad3858f187f09376eae2c8d0348e8684da07601b55259c9a8f5516984df27076458abd369b51e9745887ef6176a4fc142589dd6b4c6a2ba76498432b1d56c8a8cb746e1b46499204e37aa8d2e151915aaca7ebcfdad29e7c38c50bd1885f92db62aebe209f3739febadfca564abc740532495f6cc24a7efab38c0076ed6fcdfa575413058d7b41b7ae0d3daf072d29160b6ba9672ee00ace724a750652f42c3ebe720e860722ef6edf4c90cbf04c79f0229522307c7fa16d0b28d35614644261f4c4087e59f30d87762b98630daa700f2b7ed474cd8fecc0a934853e0d697eb9c4d5492699e32bf3bba842ecc836bab9012f5f8ada6f2aea876666d5e3b99217473c86643845277d34f6e66bdc7031e3c1ba39bc274023ec46a12fdc4af03184267e61a020ca140013d11603d94a4ae2b9d05299ca63fe2147cf1186014f9fe2fa36432581826e8707bb83c819f34ba2aea5eebfcf384612e65ca77f3d6618d1c22b32bdf20084c7378aaaae9ec3adadd0f817566c92f952f25759fdbd9b9fab13d4de1d01d008ea4a82fff700d56d3c3784c875c701629e0988f78f9b4da23ee38c127891ff6afcd1baa1b642be15976868e453703b7700ec00b8d0179c668bf6efbb29e6fb3600b5afa6a9744e0e77a54908984591636843478f793962c74261929756110f11079296fe5e207bccb728220c759815d6629b8e852d143a40ad3669fc87e15d0c3e9cfa216f42bf32eac7e5bb868de10ccac82b813efb113c6ca9859dc6185d2f8f4c3e32e02bb923f2faf2413e8c09a146802f626ccfa1890aa51d09f7ebf1c388d3364c738c0001e7ea7b98a3c585f3d5fcc400e99443c4f8db079775af63ae31f4af2cdb0ec4eaf3d9224787e38f9a131c4e23f7a00e39f2378e2eb2a7560652dae1b9d0a1baafc534f0402a872d786c838eae1d4dc97d39b0a5a00ebf59064519b25bd8e02ca74ee6ce5bed4b307ca2c2dd8b2916d49a78462688838aaa3932e0b756028d8ca299d4565c7e7464f6e98db5269319424966ccb84e96138748127681f797a76379d2ad0afac5371cdd8352079c51524cb98a2936a914988883b6af2b5976798d2b63435999780783cb17666f5036d4852126afcf64157a9b6882f6618e8ee96b2067eb7dfa6a56051ed9c3d27d3f4aa973e8c74419da22b8fc5deec21b2aa28a5b8a3a05a7e36715dc4bee3a0b302c1e17492470e0d0e9ed0328cce230dd59c52786b314c3beb5d250f4d3238b76cb3f4911723bc36752f890aee53252fc41eaf1f3c38a27f916a5784e22eb67ba1f419daafb7816c1acf6b47f280656c710986b2cdf9f646f46066beca2c032993ff8fd6a79188b4eeecff7b42c6c8b29edc3826a705e7cf51683fc2b893ab4149e2d38b56db2735e4e7c4f406c8cac2668867f80ea942e16977625529a6d1092e449a25c5ec0838cc5a119af3df3bedf0f68e421f20e8a194f108870b667fc7012e0182253fe837f743d8480c70c19c585e3fcc66fb868aaa6dd04acca85dac8f1b0ae81f983d40c79670cb070c2309dba16609c0ad4bab1f0f81b97eb73f02f65e33792f3368c89417454784610a94def5f09de2fdff66db1f5b7cc534d212bee9532d10b50dbdc1ce5629d330504b730ef324303b37a4f70f33e10de479f229ecaf59610cd8213ebda2e4b846e5ff3ba20298726931cdb47a0cbb14329f0eb0bd5904633771514d3f437609cce9b405d29ebd78cc0a06edd89e214034f632b3826b6b9f89f10234b1dd5d70c420a56cf4ef26ca5aeddfb8bc6e9d2cf1aa9faf0e5c030268aa2ed709ebb3e05f546f1be8c98882abaadcebc4243db8dc91dd4b3a9e02b487b9214c3884269270c6a8afdb479ed78b07cf7e46a271136eda03207ce49c87fbe4e7c31efef7a2dc11c822c6a835c8cf4490b2857eb5fb80d1483ea74a18ad2b4e8523a74aecac344e96abc90467aeb8daf940d0c8cbbe897467dd21085b8ae93f471091978ad2de2db0b970ffd1e4e05ba31e13abf266fb0454f60a4d7fa881c8d9c076d502cd3f97142acfa552f9bb8b0ea29c7251c85909e8c12d2694df8066305c84ac9989474064ab0325b445af736cbe623e80a8428456780a63a5e135028f2666d37c847554ca51629197944209927c9c46bf7c374fd29986cf8eee5bb6f8af1404511d925e2ba925b119fcf8afae8f4dc8b5efe378e85eabcf611a1b6099b9eb4f7704718f20a9835145189c470a998c0b4cca3dc2109b73569e79fec0ecb7433ab939ee31408b9874f345daed4e6f257428ae83e5401cbcfc7e21c0f800ead4207285376ec23eccf9694995dcd3763c2e56432bfb5fc4ad5b29a796f0d23d1918d53c8015ccb60cfd7f571173d9b4e0dc6423308abcdbb4591f6a123123498eeea2fe8a04e0066c140e2ce42ce7e9422043c9fc329d625e48c49662841f15e7d406b2d2fdd8eeb1bb1982cef7415fb80171de5b59ff5dfe303679435d352b148c11b1089df7556b2ea3e118d9f98e5a47e1510e677ea700cc02946c6338401f710f0662268b84ac1f41f820297b62abdf2893ac910ed9023a14516e49ecddd2b3c272c9a78ecafd69483d24225e62481c8fac0ab442b314eef0df19ae8bbc66e97c10fad33f057aaa20f967de79a925d067aad48dd04ee79b775beff9c4585ae96477f2e30138d6458f29ccade3c51f11d94f2704ba0cdccfade0f2c7191c0a62c0c719d4ab2b0c4c22a4af666f18ef8a3c3416834649aaa9857b4bf936f8b3e88f304508ad8373fa11deaea2f3a8c834da23cfd233874074c1396b69508f243edc880892de6676c5e78d616bc4ffeea39280373adb09f71d4602c21607d2f732e44ffc1c03704dd01526d4f67b48c706f30cc55d83f192eac735b9e29a70536d1510f812924e4c63c5ce1f7b05d4c2613968104af7b8bb30eec25e5076001f35686fd4114c11feb67fc3c4caf85f1c636c04f7101f893ea34a7d3fb348c545c4910500c515b554ac8332d8ac1b18032665b350dd7cb5c509987cb0507d9dbd86fe0a090026f5f00", 0x1000, 0x20000, &(0x7f0000001180)=@full={0xb, @dev={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @default, 0x3, [@bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @bcast, @null, @null]}, 0x40) (async) getsockname(r0, &(0x7f0000000040)=@pppol2tpv3={0x18, 0x1, {0x0, 0xffffffffffffffff, {0x2, 0x0, @multicast1}}}, &(0x7f00000000c0)=0x80) 00:08:45 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket$netlink(0x10, 0x3, 0x8) setsockopt$netlink_NETLINK_TX_RING(r0, 0x10e, 0x7, &(0x7f0000000000)={0x800, 0x1, 0xfffffff8}, 0x10) 00:08:45 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0x0, 0x0, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) (async) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = socket$netlink(0x10, 0x3, 0x8) setsockopt$netlink_NETLINK_TX_RING(r0, 0x10e, 0x7, &(0x7f0000000000)={0x800, 0x1, 0xfffffff8}, 0x10) 00:08:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = socket(0x0, 0x1, 0xfbfc) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r0, 0x10e, 0x8, 0x0, 0x63) getsockopt$PNPIPE_IFINDEX(r0, 0x113, 0x2, &(0x7f0000000000), &(0x7f0000000040)=0x4) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 1: r0 = socket(0x0, 0x1, 0xfbfc) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r0, 0x10e, 0x8, 0x0, 0x63) getsockopt$PNPIPE_IFINDEX(r0, 0x113, 0x2, &(0x7f0000000000), &(0x7f0000000040)=0x4) 00:08:45 executing program 0: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, &(0x7f0000000700)=[0xffffffffffffffff, 0xffffffffffffffff, 0x1], &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 525.681178][T29250] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:45 executing program 1: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='ufs\x00', 0x1) r1 = openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r2 = fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000100)='}:!+}.].\x00', 0x0, r1) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='-#[)+\x00', 0x0, r1) r3 = fsopen(&(0x7f0000000000)='cgroup2\x00', 0x0) pipe2(&(0x7f00000000c0)={0xffffffffffffffff}, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r3, 0x4, &(0x7f0000000080)='\xf7\xca\xf6\xce\x8d\x01\xad\x04\x98\xaa\xeb!!}.\x00', &(0x7f0000000100)='./file0\x00', r4) getsockopt$rose(r4, 0x104, 0x2, &(0x7f0000000140), &(0x7f0000000180)=0x4) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 3: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x3, 0x0, &(0x7f0000000740)=[{0x4, 0x1, 0x8, 0xa}, {0x5, 0x5, 0xe, 0x8}, {0x5, 0x2, 0x9, 0x7}], 0x10, 0x6}, 0x90) 00:08:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='ufs\x00', 0x1) r1 = openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r2 = fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000100)='}:!+}.].\x00', 0x0, r1) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='-#[)+\x00', 0x0, r1) r3 = fsopen(&(0x7f0000000000)='cgroup2\x00', 0x0) pipe2(&(0x7f00000000c0)={0xffffffffffffffff}, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r3, 0x4, &(0x7f0000000080)='\xf7\xca\xf6\xce\x8d\x01\xad\x04\x98\xaa\xeb!!}.\x00', &(0x7f0000000100)='./file0\x00', r4) getsockopt$rose(r4, 0x104, 0x2, &(0x7f0000000140), &(0x7f0000000180)=0x4) [ 525.799912][T29271] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) (async) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='ufs\x00', 0x1) r1 = openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) (async) r2 = fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000100)='}:!+}.].\x00', 0x0, r1) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='-#[)+\x00', 0x0, r1) r3 = fsopen(&(0x7f0000000000)='cgroup2\x00', 0x0) pipe2(&(0x7f00000000c0), 0x0) (async) pipe2(&(0x7f00000000c0)={0xffffffffffffffff}, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r3, 0x4, &(0x7f0000000080)='\xf7\xca\xf6\xce\x8d\x01\xad\x04\x98\xaa\xeb!!}.\x00', &(0x7f0000000100)='./file0\x00', r4) getsockopt$rose(r4, 0x104, 0x2, &(0x7f0000000140), &(0x7f0000000180)=0x4) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:45 executing program 1: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x25, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10, 0x6}, 0x90) 00:08:45 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 525.920442][T29290] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:45 executing program 1: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='ufs\x00', 0x1) r1 = openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r2 = fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000100)='}:!+}.].\x00', 0x0, r1) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='-#[)+\x00', 0x0, r1) r3 = fsopen(&(0x7f0000000000)='cgroup2\x00', 0x0) pipe2(&(0x7f00000000c0)={0xffffffffffffffff}, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r3, 0x4, &(0x7f0000000080)='\xf7\xca\xf6\xce\x8d\x01\xad\x04\x98\xaa\xeb!!}.\x00', &(0x7f0000000100)='./file0\x00', r4) getsockopt$rose(r4, 0x104, 0x2, &(0x7f0000000140), &(0x7f0000000180)=0x4) 00:08:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1}, 0x4000000) 00:08:45 executing program 4: bpf$PROG_LOAD_XDP(0x5, &(0x7f0000000780)={0x6, 0xf, &(0x7f0000000400)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, &(0x7f00000005c0)='syzkaller\x00'}, 0x90) 00:08:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 526.002518][T29305] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:45 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='ufs\x00', 0x1) r1 = openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r2 = fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000100)='}:!+}.].\x00', 0x0, r1) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='-#[)+\x00', 0x0, r1) r3 = fsopen(&(0x7f0000000000)='cgroup2\x00', 0x0) pipe2(&(0x7f00000000c0)={0xffffffffffffffff}, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r3, 0x4, &(0x7f0000000080)='\xf7\xca\xf6\xce\x8d\x01\xad\x04\x98\xaa\xeb!!}.\x00', &(0x7f0000000100)='./file0\x00', r4) getsockopt$rose(r4, 0x104, 0x2, &(0x7f0000000140), &(0x7f0000000180)=0x4) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1}, 0x0) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1}, 0x4000000) 00:08:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 526.116149][T29322] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:46 executing program 0: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) fsopen(&(0x7f0000000000)='ufs\x00', 0x1) openat$vimc0(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) 00:08:46 executing program 2: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) fsopen(&(0x7f0000000000)='ufs\x00', 0x1) fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: r0 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='\x00', 0x0, r0) 00:08:46 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xe2) 00:08:46 executing program 4: setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(0xffffffffffffffff, 0x10e, 0x8, 0x0, 0x0) fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) 00:08:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 1: fsopen(&(0x7f00000000c0)='hfs\x00', 0x0) 00:08:46 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 3: r0 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='\x00', 0x0, r0) (async) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='\x00', 0x0, r0) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 2: fsopen(0x0, 0x0) 00:08:46 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x108) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 4: fsopen(0x0, 0x0) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: fsopen(0x0, 0x0) 00:08:46 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x10a) 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 4: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) (async) r0 = accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000000)) 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 2: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:46 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:46 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x145) 00:08:46 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: r0 = accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) getsockname(r0, &(0x7f00000000c0)=@tipc=@id, &(0x7f0000000080)=0x80) 00:08:47 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 2: r0 = accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) getsockname(r0, &(0x7f00000000c0)=@tipc=@id, &(0x7f0000000080)=0x80) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) (async) getsockname(r0, &(0x7f00000000c0)=@tipc=@id, &(0x7f0000000080)=0x80) (async) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x145) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$nl_generic(0x10, 0x3, 0x10) recvmsg(r0, &(0x7f0000000200)={&(0x7f0000000000)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @mcast1}}}, 0x80, &(0x7f00000000c0)=[{&(0x7f0000000080)=""/15, 0xf}], 0x1, &(0x7f0000000100)=""/237, 0xed}, 0x100) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:47 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x20e) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@short={0xb, @remote, @null, 0x1, @netrom}, &(0x7f0000000040)=0x1c, 0x0) 00:08:47 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x300) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 0: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@short={0xb, @remote, @null, 0x1, @netrom}, &(0x7f0000000040)=0x1c, 0x0) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) (async) syz_init_net_socket$rose(0xb, 0x5, 0x0) (async) accept4$rose(r0, &(0x7f0000000000)=@short={0xb, @remote, @null, 0x1, @netrom}, &(0x7f0000000040)=0x1c, 0x0) (async) 00:08:47 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x500) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, 0x0, 0x0, 0x800) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x500) 00:08:47 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x600) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xfc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xea, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xfc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, 0x0, 0x0, 0x800) 00:08:47 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x700) 00:08:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:47 executing program 4: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x800) 00:08:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 0: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x800) 00:08:48 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x801) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x94, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x81, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c"]}]}, 0x94}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$nl_generic(0x10, 0x3, 0x10) recvmsg(r0, &(0x7f0000000200)={&(0x7f0000000000)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @mcast1}}}, 0x80, &(0x7f00000000c0)=[{&(0x7f0000000080)=""/15, 0xf}], 0x1, &(0x7f0000000100)=""/237, 0xed}, 0x100) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:48 executing program 4: recvmsg$kcm(0xffffffffffffffff, &(0x7f0000001540)={0x0, 0x0, &(0x7f00000013c0)=[{&(0x7f0000000100)=""/81, 0x51}, {&(0x7f0000000180)=""/89, 0x59}, {&(0x7f0000000200)=""/4096, 0x1000}, {&(0x7f0000001200)=""/189, 0xbd}, {&(0x7f00000012c0)=""/231, 0xe7}], 0x5, &(0x7f0000001440)=""/248, 0xf8}, 0x10001) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) r0 = seccomp$SECCOMP_SET_MODE_FILTER_LISTENER(0x1, 0xa, &(0x7f0000000040)={0x4, &(0x7f0000000000)=[{0x1f, 0x84, 0x40, 0x400}, {0x100, 0x2, 0x7f, 0x6}, {0x81, 0xf8, 0x1, 0xfcc}, {0x2, 0x5, 0x2, 0x2}]}) ioctl$SECCOMP_IOCTL_NOTIF_RECV(r0, 0xc0502100, &(0x7f0000000080)) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb5", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: recvmsg$kcm(0xffffffffffffffff, &(0x7f0000001540)={0x0, 0x0, &(0x7f00000013c0)=[{&(0x7f0000000100)=""/81, 0x51}, {&(0x7f0000000180)=""/89, 0x59}, {&(0x7f0000000200)=""/4096, 0x1000}, {&(0x7f0000001200)=""/189, 0xbd}, {&(0x7f00000012c0)=""/231, 0xe7}], 0x5, &(0x7f0000001440)=""/248, 0xf8}, 0x10001) (async) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) r0 = seccomp$SECCOMP_SET_MODE_FILTER_LISTENER(0x1, 0xa, &(0x7f0000000040)={0x4, &(0x7f0000000000)=[{0x1f, 0x84, 0x40, 0x400}, {0x100, 0x2, 0x7f, 0x6}, {0x81, 0xf8, 0x1, 0xfcc}, {0x2, 0x5, 0x2, 0x2}]}) ioctl$SECCOMP_IOCTL_NOTIF_RECV(r0, 0xc0502100, &(0x7f0000000080)) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x900) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 2: syz_init_net_socket$netrom(0x6, 0x5, 0x0) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xa00) 00:08:48 executing program 3: syz_init_net_socket$netrom(0x6, 0x5, 0x0) (async) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 1: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:48 executing program 1: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 2: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xff, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:49 executing program 3: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xa01) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 0: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x58, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x47, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa3127"]}]}, 0x58}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x1c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xc, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic]}]}, 0x1c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: set_mempolicy(0x0, &(0x7f0000000000), 0x8000000000000000) 00:08:49 executing program 2: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x58, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x47, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa3127"]}]}, 0x58}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 529.361530][T29851] __nla_validate_parse: 14 callbacks suppressed [ 529.361544][T29851] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:49 executing program 3: set_mempolicy(0x0, &(0x7f0000000000), 0x0) 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 4: set_mempolicy(0x0, 0x0, 0x8000000000000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x58, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x47, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa3127"]}]}, 0x58}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: set_mempolicy(0x0, 0x0, 0x0) [ 529.547298][T29874] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 529.589959][T29888] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) (async) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) (async) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:49 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xb00) [ 529.639208][T29899] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:49 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x74, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x64, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda"]}]}, 0x74}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 4: set_mempolicy(0x0, &(0x7f0000000000)=0xffffffff, 0x8000000000000000) (async) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:49 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$netlink(0x10, 0x3, 0xc) getsockname(r0, &(0x7f0000000080)=@pptp={0x18, 0x2, {0x0, @initdev}}, &(0x7f0000000100)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) 00:08:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xc00) 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x74, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x64, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda"]}]}, 0x74}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 529.828030][T29930] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x74, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x64, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda"]}]}, 0x74}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$netlink(0x10, 0x3, 0xc) getsockname(r0, &(0x7f0000000080)=@pptp={0x18, 0x2, {0x0, @initdev}}, &(0x7f0000000100)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) socket$netlink(0x10, 0x3, 0xc) (async) getsockname(r0, &(0x7f0000000080)=@pptp={0x18, 0x2, {0x0, @initdev}}, &(0x7f0000000100)=0x80) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) (async) 00:08:49 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xd00) 00:08:49 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x2, 0x0, &(0x7f00000000c0)) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x73, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecd"]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x100, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xee, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x100}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 530.125627][T30005] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_rx_wq": -EINTR 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4, 0xd3}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xe00) 00:08:50 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x73, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecd"]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 530.155402][T29978] workqueue: Failed to create a rescuer kthread for wq "nfc5_nci_cmd_wq": -EINTR [ 530.236750][T30020] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) r2 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r2, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x73, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecd"]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xe02) 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0xeb, 0x0, 0x0, @u32=0x2}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 530.653050][T30097] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xe02) 00:08:50 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(0xffffffffffffffff, 0x5, &(0x7f0000000080)='\xec', 0x0, r1) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7e, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d6"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x8c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7a, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295"]}]}, 0x8c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7e, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d6"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) r1 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r1, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7e, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d6"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7e, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d6"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) pipe2(&(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}, 0x0) r1 = openat2(0xffffffffffffff9c, &(0x7f0000000540)='./file0\x00', &(0x7f0000000580)={0x40}, 0x18) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(r1, 0xc018937b, 0x0) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000080)='/dev/virtual_nci\x00', 0x0, r1) lsetxattr$smack_xattr_label(&(0x7f00000000c0)='./file0\x00', &(0x7f0000000100)='security.SMACK64IPOUT\x00', &(0x7f0000000140), 0x1, 0x2) 00:08:51 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0xf00) [ 531.178313][T30201] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.0'. 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @remote, @netrom, 0x0, [@rose, @remote, @bcast, @default, @bcast, @null]}, &(0x7f0000000040)=0x40, 0x0) r1 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r1, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 531.252748][T30218] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 531.296112][T30213] netlink: 116 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:51 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32=0x2}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) pipe2(&(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}, 0x0) (async, rerun: 32) r1 = openat2(0xffffffffffffff9c, &(0x7f0000000540)='./file0\x00', &(0x7f0000000580)={0x40}, 0x18) (rerun: 32) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(r1, 0xc018937b, 0x0) (async, rerun: 64) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000080)='/dev/virtual_nci\x00', 0x0, r1) (rerun: 64) lsetxattr$smack_xattr_label(&(0x7f00000000c0)='./file0\x00', &(0x7f0000000100)='security.SMACK64IPOUT\x00', &(0x7f0000000140), 0x1, 0x2) 00:08:51 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x1100) 00:08:51 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000000140), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) r0 = openat$cgroup_devices(0xffffffffffffffff, &(0x7f0000000000)='devices.allow\x00', 0x2, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='devices.allow\x00', &(0x7f0000000100)='./file0\x00', 0xffffffffffffff9c) write$cgroup_devices(r0, &(0x7f0000000080)=ANY=[@ANYBLOB="f40c62e54d517c68a7132eb5169a040085dc53f77d9b7b6b5cc55ea443c0faa9ab8becb56078458f3820284567a95b6f3420e2a914143b25b6a333c9a0e1d2b7df1885de4e23b65aab6275744be57ef6d65cdc9899ca8095218d34afd8db945949e3729a6a91a93513111504f2afabaf2f04e9b29eeea7eda5b51f019ee8"], 0x9) 00:08:51 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x7f, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x1200) 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 531.740944][T22946] Bluetooth: hci4: unexpected cc 0x0c03 length: 249 > 1 [ 531.744428][T22946] Bluetooth: hci4: unexpected cc 0x1003 length: 249 > 9 [ 531.748219][T22946] Bluetooth: hci4: unexpected cc 0x1001 length: 249 > 9 [ 531.750692][T22946] Bluetooth: hci4: unexpected cc 0x0c23 length: 249 > 4 [ 531.753086][T22946] Bluetooth: hci4: unexpected cc 0x0c25 length: 249 > 3 [ 531.755193][T22946] Bluetooth: hci4: unexpected cc 0x0c38 length: 249 > 2 00:08:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x1696) 00:08:51 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000000140), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) r0 = openat$cgroup_devices(0xffffffffffffffff, &(0x7f0000000000)='devices.allow\x00', 0x2, 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='devices.allow\x00', &(0x7f0000000100)='./file0\x00', 0xffffffffffffff9c) write$cgroup_devices(r0, &(0x7f0000000080)=ANY=[@ANYBLOB="f40c62e54d517c68a7132eb5169a040085dc53f77d9b7b6b5cc55ea443c0faa9ab8becb56078458f3820284567a95b6f3420e2a914143b25b6a333c9a0e1d2b7df1885de4e23b65aab6275744be57ef6d65cdc9899ca8095218d34afd8db945949e3729a6a91a93513111504f2afabaf2f04e9b29eeea7eda5b51f019ee8"], 0x9) openat$nci(0xffffffffffffff9c, &(0x7f0000000140), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) (async) openat$cgroup_devices(0xffffffffffffffff, &(0x7f0000000000)='devices.allow\x00', 0x2, 0x0) (async) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='devices.allow\x00', &(0x7f0000000100)='./file0\x00', 0xffffffffffffff9c) (async) write$cgroup_devices(r0, &(0x7f0000000080)=ANY=[@ANYBLOB="f40c62e54d517c68a7132eb5169a040085dc53f77d9b7b6b5cc55ea443c0faa9ab8becb56078458f3820284567a95b6f3420e2a914143b25b6a333c9a0e1d2b7df1885de4e23b65aab6275744be57ef6d65cdc9899ca8095218d34afd8db945949e3729a6a91a93513111504f2afabaf2f04e9b29eeea7eda5b51f019ee8"], 0x9) (async) 00:08:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:51 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x88, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x78, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x88}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x16cd) 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.038284][T30320] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_rx_wq": -EINTR 00:08:52 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x84, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x74, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic, @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x84}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.044328][T30333] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_rx_wq": -EINTR [ 532.088408][T30288] chnl_net:caif_netlink_parms(): no params data found 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 1: r0 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) r2 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) r3 = openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) r5 = syz_open_dev$ptys(0xc, 0x3, 0x0) r6 = openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2", 0xffffffffffffffff, 0xffffffffffffffff}) r8 = openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r9 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r9, 0x0, 0x0, 0x0, 0x0, 0x0) r10 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r10, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) r11 = bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) ioctl$SW_SYNC_IOC_CREATE_FENCE(0xffffffffffffffff, 0xc0285700, &(0x7f0000002f80)={0x10000, "deb61e2f2a84540b872481f48e71267ebf181545a2e58b1a73b1d5c64fa91704", 0xffffffffffffffff}) r13 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000003040)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x101, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x0, 0x3}, 0x48) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000002fc0)={{0x1, 0x1, 0x18, 0xffffffffffffffff, {0xee01, 0xffffffffffffffff}}, './file0\x00'}) r15 = openat$autofs(0xffffffffffffff9c, &(0x7f00000030c0), 0x2c040, 0x0) r16 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r16, &(0x7f00000046c0)={0x0, 0x0, &(0x7f0000004600)=[{0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}], 0x9}, 0x0) r17 = socket$nl_generic(0x10, 0x3, 0x10) r18 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000003100)='./binderfs/binder-control\x00', 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f0000003200)={&(0x7f0000000100)=@proc={0x10, 0x0, 0x25dfdbfe, 0x20000000}, 0xc, &(0x7f0000001800)=[{&(0x7f0000003240)=ANY=[@ANYBLOB="480100003c00020029bd7000fedbdf25c80059800800b60005000000fcef61d3255f79e749b74989cefd2d67f4a3a7f5d872130e436ace7f0251338b726457f2101bcb86d236c5477fb7149c43d955ff467b1d1d809fe6128edcf1b58e28e82300c5c2573e2dc0ad416e2f2871b5d459581f24b06e903ecd8649a80346f46972ab1e0e864198d13853cee89e3a595105e83061e0793077e3f466228c08009600710000000800af00", @ANYRES32=0x0, @ANYBLOB="0d00660080000000000000000500100000000000150098002f6465762f7669727475616c5f6e6369000000000c00c800b8000000000000000c0005000100000000000000c866ef508d669a67a13484453afb67d337f7230ab1bd797f79bcaa1764573b3e11364e29f3db8e20a98ed041912cdf40a5c23af50e2e3a7d2e9315821ce9e21b034756e42c2eb5d5f961a7cae21fbc08bed3e4ad7e2847c6834e35f695631a9b6f03070000"], 0x148}, {&(0x7f0000000140)=ANY=[@ANYBLOB="18ff20471b34cc000000200200022dc77000fedbdf250800e4065f3a788261d9ff0f971ca11525fbe80a9819385c78253494efe3e46d60cc422d35bb9907341cc1a20139fdeb7fbcb083617b8a49cc448c78ba24e08b35f47df84e534deb9226311e8b2aaef55c3a6d1987a75e03ffd476638d92a519543bcff12c98ad38da7835761dd0cdeba44f4d8e9c00000000e64daefac1c90189ce5bd8cca7dfc5689f5c6e6e3cc610249e972f824fade4f3920c46ebb50e3fc50791edff0fcc52b365e1da600d02eb7f000000"], 0x18}, {&(0x7f0000000400)={0x20c, 0x10, 0x100, 0x70bd25, 0x25dfdbfe, "", [@nested={0x40, 0x131, 0x0, 0x1, [@typed={0x8, 0xf2, 0x0, 0x0, @pid=0xffffffffffffffff}, @typed={0x8, 0x9a, 0x0, 0x0, @pid}, @typed={0x15, 0x27, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @typed={0x14, 0x58, 0x0, 0x0, @ipv6=@local}]}, @typed={0x4, 0x64}, @generic="c789e7cc7ca8dab63be9ed69e6e9a202e2d52f41aa", @generic="bfa355ce6fab563c093e78c05135747bfff952128f788a9c1ac249a7191e52772e6dabd2567244773291d880df05c45846fb4f3f987662064e2bb43b7ed60d3c4b6e54c4409c3ed0d40e512cbb060d95117797bd7f17b73f40f7dc49ea6c0be1efabf891a76d6f8550dd5d4686af8f9075a2c165e3a108c2c6f279672edfa2e65ce7a96a1ae680700aa7fe2a58ce", @nested={0x114, 0x105, 0x0, 0x1, [@generic="924193881425028ded5e45af38309971b787cba9b5eea3c02361f21ad6ec6bec8f6a34d2be156296c42513da5533fb9d3af4038f7b02c9fd412b3b70f94bb65e569922e3e9f346c0b8ed2ad109e0d5f3b953e5411ddfa7ed7b0a60542e24b0510f3d1a96629ba0a58021eb623f3acd60ad8f79c4278dd5c9213020d6359fb392fec783bed11fc3484972230b510f0e63cadf00b0b65fa135c61fe0cbf3280ed4a745719d8945f1fa1864a507ff66b293b5c51bdcb35a527746d528fb38c049829187d0a677b124fa68b2d6ceb565e1321b86fc", @generic="4d721c8b383769a764937f17db741ac512ad5f84dc682af8edeff4a74f6090876cc97ba3b128eecb2eef11038b50a328d818c4776ad01890bf08440a9a"]}]}, 0x20c}, {&(0x7f0000000740)={0x1084, 0x10, 0x800, 0x70bd25, 0x25dfdbfd, "", [@typed={0x14, 0x5d, 0x0, 0x0, @ipv6=@loopback}, @typed={0x24, 0x18, 0x0, 0x0, @binary="17089d9abf899777e6a2d77f1e7d5724872ee91444fa4da34531d08dc7f9ae85"}, @typed={0xc, 0x128, 0x0, 0x0, @u64=0x9}, @typed={0x8, 0xe9, 0x0, 0x0, @pid}, @typed={0x1004, 0x25, 0x0, 0x0, @binary="4d25059a1a01986681c822bc09f098bb73c119783f81dbe8824a8fa9014888c3f670af704119f6cc1545d35174c987fb52d9051bf3b01e4c7a424f595280a47cc3a01b45f06cba1defe1ab7ce1d9f66c00f7e039720b0dfc9e1d53cb304618a32f26256dccd3da1d40e44d66bf9f3c413ebe925e39394fb082f7818758d0b7e58df5069d078aaf74adea049747759263eea91b63f86a6ddf0b3290685b897c6073700c5c1dd8fea2dd1d01602e4299d5f125cf0163fa2c1ebc6de3f042eb9bde2e479ebd7a738f35c4cdb587bef8e571e0edaf3c4705553180e1cf9bf0adf27fd2835864e25fad08d8cfaac58e6c60d720aa7903db9ccb41fcf8793e14f97c81867739fb11c90d583f0a35654979a5a00e8e43e542077da3c7d6f7466d5a00d5d60a23b518e501fcd3e489c97762f6173cbb973c450b3b26edeb7a730407c74e7f11aa25de3bfa907086945e67b54a837c73567ae62633dafebad4d38ab5813bdd9d80262b6d2471a1a61cb97dc4fcdebf0cd980c04319b0463b1d0db04131c411e535438233a2028650ace7688bec3de1806e44ea79a63aade453acfcdbf5aa3e63c3e075a5caf2bc745d88554b7d23afd40094c9e654a253050ccc8e1d549c6042c9f8da7cbea3898c8573b04b30a678bf6d39c0f68fd2cded90da3dc4a008dcb455a30765aba41fdf7b9324d7ba34ac8dbbd47bdbd1b022b78ce2b765342dd776059d3a79d86b317e199147b1933b0b1def209de8492bd6962fbc93338eea9a5b3288161294459c765bb8a982f8b16f258aec9efbaaf651667758f05a7e11163dcd36c75bccc3a6df491499ec6b5c9e86dbb37a495033bcf5fa1a4ab5de98a8810e717c7f8db4a22076ddc32271f4fcf615c0a80655a7bb304f1d020587f34e6e7c1b99b6102ef51a8d92d7373d849e170129f4a1801d45d5aaff670c07afddfdc55fb0a597951c58603585f247f5a11f9c35ec83a1680f103f8ecf6deead4ad441cd67a4049486a16ccb6ccb5e935bf3a002312a2c419a0b14dd436730bd9b7f5c74c36f6b3dbcd184d750d0278fb1f8184d0e7545b7d7af48d71ca0372ea5b8e9d8ad797dafce2418bb24513a2653a7786998669795d8146f5b7ce28fc183aada6a942d1e0df98c1dedb579a46ed84a97798e6c4415266c114c659b01cb35c88a6c4b4b0383c5fc2b06c551319ebc9040a9d2939cc8bc1f1e5d0e156ef1d4d782ed7d87a28f3ef15a4494ef55ec26eb665078fa23548b9a66e438c3f747ff20ab2ea6aa03298b45fac67ff91dc3aee6abc127d5ddfd1a6f8d91999f85bc95483ddf64ee31f0a3d93312b5b29083233478ac85e37f9550f12cebeb072d24932145c7a211f4420f2bcabc52cb4caeb92354021181554ad0a4ae1c0d9d90707990362236d42cb29147f27d6830a4123d0ac346ee69a91922107d80b3457a66f274d98ef2d867022b34379d229ac4e702a64db3bf011bd1b4daf0a6d878ee9b681c9f52e5560a20225e871fa7ac880881f1e39a602400ecfa89665bb2558181232ded9fdf3a1063f4a1ff36241fd0ad58f09c05471dfc263a52dee36cf97f0915c22e7b22abfece46d5b0650fa1517a0fc3f820d5c7b378256daeee7723d9f04858747d630d3fbfcae75210663075676d890c492a50a0548f039deb20fe0a33017c1c721df39883dd6484267e05fb244370fcb729b522eee69ef47ec7934a380531a5f42b139822b6c7cff00d5cfb291ba9e2edcdf8b8b0735e49ccc2dabe9a1c363feeb821b04a645d3686544a212e887aba2bc4b6f171873c465a8a70e0810f036a73eb86ffc67182460373fb3a1b858eaac511934177126ac3060b11480fbc4474b558c496a27221a55d72370b498e081ed17d1197a42b373ccb86d76a2487dd66a60acc61bcbe7f61b72bd9e533793b7afc1425f1e8133c791cd65b4e766642c9497fc5ccca4a530d0235a620d2b08e2bc6e2b940d7e2b6312d489f63d525ea2e8c4df4d4ac05435b63681b91b13ae0d203eaa722662cdc33b8ac68fbaa2ff7f657952dd46650bb09633c519cc7c44126ee05babc55a3fbc2933bedf7cffcc55d7a3fcc4460e040e2896d1d8070246c821eef4e6723bc1cab07b305a4fb4ac6d78898d12144f56e25c8c5df9e6c80c7f866c3961ecbbaf1eed026da06b776c5591e1af5a7954bbb4efb0c1af303b688aff0576b1c50b79e6e310ba020ce3dcb564dd606d650a3dc2716d157bf013211468536ad0e6646213eb3fb02a884dac52b54a5aac329ec1eb4a5fb2beaa82f4a245410b3e583cd3c0014fca640d7a752ca6dd47c271d1d4ada9bb79b42501cb4f5894b9b9681a43bfae67b234288ee50e0101decda623569d5d9dad0a7e4c608858fb04de1cf356c9f0db9ea9cb5df287fb23efa0b17082c1257d77c16539bccfd3e7591540eb82e9f765dbb5bb289f2bd489a0488032b5461d1348d670693f4e53341f2456ace8e2f59ac470dc8c999e9e52c2e0e320a0e2d1238296d1185561002d93825b147935945c6d0139fe99d3a155a5c1afc2f3f176dd521fe4927737d83d33ce48d333f39158f5653f8345b7cda03ef8dd217868e7b8e06ca3b771fb923458fb649aeeca8e9f08ff2db3872ffffa6e7f773c37da792fd1c3030fe94f1ef0b971b4b6632783f15dcadfaab99be10db107957061340730f37c1c5655117dc0115f33987e051e6acffaaff9491694e4de23d08a24a986c1e866dd1910d215e22b591069971355cb145f52fc22caf8f5b120a9f4f9723cae79e1d15f05a2494d730e31e548148e9b326f37d9565157d03415b9f62c631d41c7d8d0b40b5b99f02dadc6ab95d81a97098628cf1c2803228169e4aad29f298753ef948a89c9dd38d4e953b0f3ec6d97a33d49c097baea530ad9f766d51f695976398d8e10f8c3fcf16b2bdf8f1d2535accf1fa0e85edbc4f189dc4ce16c9a7330fc77864b67a6b296b1349efc28d4281320214024fa851f467d3acb8d3b56ec98ffc73a330d8c395c5b2b264b24dcd6d03d4a7f558b30be5b22c6707b42e855a08df63825a89157bf76031f574454b142f05fe90bc9eeee2d7669615aff7233e6c01dd10da16a0fb5b488429e1fd936fc60b4897bc62c708e6df7d244e1631879bed247ae681bf434196d8db50ab1f702775c567bbc7f811fe4d00682f1d0084f67610318bee61189ac300da5cacacf85f067b5b21c6d68cdd3d11d48f365c7a187c2895765131800ae6d4b64c37ab8211f3d870284edbce98669e6721969e5690876f6807b9233784bc58a50cc75350f21622cb9e582644571add1159c307e0c2d385eef4b92faf94826b52fe7075ee40d891114e9ca69a389e1d571cd4c66bee43876b1251828a1a181b4011197c3fdd2ee610dd463320afc1e2f17c1f23ef6b4dfb1a73dc8747c94060a741cabedbad0114bfff358d0508c977b784f0d352de6b7c1e49dfe47e99211483be99e7b33147394f0fd9bc87f3cf0c02d174fbb09ff70add88ac6d5e243245b09d8e2b0d349698bd38f75ccfbd989ceb1bc4bb5732eddb177f97c80e0438d5e1ba15c9ff2ff50ba19ad93b4356a2664a73c8cb33663097887d7fe88c1de3d78309ebe7a16f483f544406ac448aa9356bd9171e4aa027e43295bc260bb1b6ba1fb00a4ac4490d1dd47e84bd6593231e1ed6e1e4d1defe56d7843e5a0120c2d33a9303349f424fc4476f27e2454edf30954496e5b4e789802eae2229f43073ad5c1ef64ffa3f8b4d0219b8812f2991866c36a7b1dfd8b510f69d1d1549d26db9d3c3cf7aefe1665c81082f2b9fe58d0e7670e9fdcda32566cb396a5ff81852b93945c6720554f36dcbc79bd44b5145d9e8382d6277afa4eb6a28d36f8239a533be4f9ee1fe8447ed63909391208b8b3af9df22e834ec9718f3474e73532fee8e113066d8eb5c317e97b56e7aa0749b4d84d8dd5435000c05e6b73ad4978cdbcc47480737923eafc754ea4587ce6b9b08269a3a73838e93c226d50449eeda62858617f98846c0658e9ffa8ecd2559b21fea3db08f413b593fa8c3e09d572993fe7acb83da43d75a5a789e5bca383f32966271ef7d4c1c6be53b98a1a7ec2c23957cccf019f7a7c0f9a87d89f1c15bcedbcef2bc4ce081aff354bc3df005ec813693348af7a4f6fa81ae4ec3616d63b07e1397cab03dc36654503b38dfc02250e00ce7e6d73ea53437b467df05979636a9d50e0b326049db5e801c4925ef646d85eb02b46d40a596b5c9b05cf31f441b97afde728d4c4c8062603f18db5b60eb327bc39dfe963601561728d4537c1da7a6120204b8b756d950c8779aec13c1891e0c1dc090857d3a5c15c074c2c86698dd9dbb43b8d789498c500058a3ac00b04ba5b4c648a7192751fb1d32d40786774ea04ad350ba349b7c0e37c39cd25d2e93d9ff06b0be3646fad982ff23f4f98fdd724396b641746ee2d8f786e7e81b16b59f4e5990ec945bc0294145e63295618e6f94ea3ab91e8bf28e65a395a27716ee8cd93d8095ee9f8472d879a383290ba00d74185bd6a34cd205eda9c4812f37eb1216cb6af28c87aec4771d7e2d9f52e525a3cc9abfc72663e5d791d4350b4393c7cb5e0572eb96de78389d794f80ebd9bc92b49445f6b68c52e6c5a6c68cfdef5b7f59626fb2a8026b906bb8c0e86f62a364c16ba1bb9ccc8f8288cc6cced7e1ecd974090dabc3a517634ce63d589b7256398037402969a485e92ff6fc6d9a572d3b9d423c1f76ac1de491792f373aaa2bc89dfb343f321b9949918504f515897f0723beadecb233327cb163116eabb36bb6543cb0cb3c2cb1145797afcfc6d626797b28f5f03bc2cd1d7f555283233d50dac1dc1f805782952e8a34c0951ae8a6691c31fa1eac2bed1c33c8ee5596002b3c687ebcee2d47475276d1cc1627866b904f5c3330fdc47061b7e104b4c59a5d019317865e71c347661f0aaa5035cfb216bcc3427fc0cfe06fc8522f04e7118584183e8722dfea9bb629a365d0b77e0e4689598a407765da16483309770b1186262f9625d507f93da36f2cda11c97c498a3cff11d6f351346f882063a277acf0f80bd452e90a44c161893b4aaade86750aa6c1ddd6df3254e1dd4e3a18054a369fc9e98ebd81d7ec35bfecda1a6ab66ec8c63331369649446d9d0da3ab9c288b83c0db15863c7aabcc68c4249b8cb7d12b46d9baeb2752ed26d71ffe48ad6392693153b13e7c8cc326c12b5603cfc9f2172269f547d2529fc70291bb38e3bb3279b9d765969bbb62b0b905251de86f0032ec26b502e907ca47756acb5bf49ac9014690d0a112d5bcf748a0fc1a6a65e9b69280ace21e9b048d4c519f0bdfa4639c5075e1cd724f3483056bca45f273ebc304ad2c26a651c31559accc7170c4f227a260fcd6127d5283dbca6241631d2752902223ad84cc715b219ccb2cf7ee1f1624fcc63739d828c543bd2ba6353f784efe7dc0e9c7837d5fdad8bc35849b18299eecb7936231181a87a3db380760c98f02389c37c0ed6758ee80d5497d0e1b2f3367efcd0225cccc3a09db7a8c8801ff8dcd86d92b53def8b3a61dc00de3f151628603238fad3b1c1cd38d98a0da9e30ab6e7228eb9970549ff06391adfa918f1257a827626b9c304d0687d0c163c3563867c317cacd371313ca74d0d05326546e9df2f22f9b1c21de253991321a13452a3b800c64e0f33c4d9b232058ba40ab92b2010464b4ca8fc10de7baf4860d80ce4b948ef95f6c684cf3c061461eaf5341c09c596eeea339e626c2f4f7db554283d590fa9b83"}, @nested={0x24, 0xdd, 0x0, 0x1, [@typed={0x8, 0xa2, 0x0, 0x0, @fd=r0}, @typed={0x15, 0x82, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}]}, 0x1084}, {&(0x7f00000019c0)={0x1424, 0x17, 0x2, 0x70bd29, 0x25dfdbfc, "", [@generic="133f5ce256e48e186b57eb92e3d67968193a2cf400d7cf3041790bce471bc257c3438d3e71a7d8393ef1b223f84de683314deeff92769a9fd838660088a17f6c2d29881a441cc1f341dc6ee6f4c5844598db7afd982706b468a39ff3ea4ad1631488fad356047ccc217c45ca7627ffc75733e22f495b0f256e7210a8d146509c7af6c38561393519a3b182b78ec1dd9bd25477a3cf587b27", @generic="e88ec576b23c16f99f3075597fe305d7cca55f465552bc35edd73b9e4d5ce2b8b486a37fc83eb8ac46f197789352d0b11938a7ba84027eff3fad4ac1222bb0d37e0471e0171b48d59270a259b2744ff58a8de6debd54f67c78bb47fbd8121844e292a0cb7e8a95c97e149a4d8f3ff033d6975eac8d84cbd3bf146fff16bc4690f3f3ac657ca1b672ece1672fbdb017c389dc634bf6bfb31190643d731343196c3facfa4d7450b8271a53706e4d46f5f873e479f3e1f3304d18c78039dec6ce750ec093f530b30b046597087ca6a4cb344ced7bdae4bbbcb05d3ef335cad3a84c072e397d382e1a4badfd816f22f0e243e0fcf79bbf3a6ce8f2", @generic="143fdd42b4d1a20607d75360dc83c395c8c3ce8252a2ce7008653d1b5e53fb041a8ccac0cb52b2df15a1e74b2e4c43a4ea411c43ffd236f7c089da121b3494d15fb915c4a8f58de93e5a92789bc38b542602a9d8af3e242d4a2e", @typed={0x14, 0x7, 0x0, 0x0, @ipv6=@ipv4={'\x00', '\xff\xff', @multicast1}}, @nested={0x120, 0xc6, 0x0, 0x1, [@typed={0x8, 0x3a, 0x0, 0x0, @ipv4=@initdev={0xac, 0x1e, 0x1, 0x0}}, @typed={0x8, 0x75, 0x0, 0x0, @u32=0x401}, @generic="b6895074bb7a3afe9630c6f99260b1f06c93bc2c0722540649440807d7a24b97f13ff6d21106215013a3fd5699a435b16158e53afb0f6ae0ff043d6639be1fb09a7c68", @generic="2749fb2104172d73bb85cc455515bcaf861c1add239821475b56e3cb827b64c9e8c2862d0fe2709bf0adc4377a55e79ee35151f9038a77fd6c62c16c9c743284915f107162cfc72596d29436f2527a36488416958a5c68125d51c5144522d271b74c6073546aae75103ccc5369d329bc92d13548f43eb3ff88fb916774b5d66961283b0fbbb72316c3552ef5507be8157459bd55be305195f50488b1a1", @typed={0x8, 0x105, 0x0, 0x0, @ipv4=@multicast2}, @generic="fd03adb6efa816bd7ffbfd7b56324c6291c0d47943cb464325b3037bcd446f2e310f70b6"]}, @nested={0x10da, 0xd, 0x0, 0x1, [@generic="845dde559616927f6d889da26095d1f40f2712fc6c56ebc6a455054ab7", @typed={0x8, 0x82, 0x0, 0x0, @u32=0x3f}, @typed={0x4, 0x50}, @typed={0x8, 0x57, 0x0, 0x0, @u32}, @typed={0x4, 0xa5}, @typed={0x15, 0x10b, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ed4141bc4f65989a9acee3dac370708ba95194311416f97e58f4a032c20ce1e20cdda03f81f0af7eed5353f09df66fda0010f3adb90b890776cd0a4e3436e5be3ca67c868e0ac0d2afe1832180a82f908afcbf870998621b", @generic="035ae798dea74a23cca6ba07b89ba8e528f72a786258952651b2346be481a4d62a955679adbf5cb2c0a16f7a01a314546c02d732e756dfee390855b0d51d78a4f4725565deafbcecbb65ab1fa666fef767d6019ed11646ccc00c84e249142e19a026cfaad65b4d2180b754ab21aa2d8c6c89eb9b2de191e769a05af2d8dc62f9bc80a546c4dd13b1f23f75253431fe2e92c09a577c7ea70dd8d9a348ecf6e7ccbea1c0c67ff07ba9edf44e3b3e50cc631d175fbb1480363d63c066920b56209b292fd82d1f11c66b7155ac3740daafd76cb43e70aacbcf63ce4e63604b6f9b178a6bf3dd0ebe5261ecb144838673d88e77eaa8686a4f9bb7eee95ed1e5698275ba14d32366d90e12551165017478e4774d497dba0995f4aa962010697dc9d7743991ca5334f52acb6f46f1e76a6c7d353d9f4b2de20ee797c17c599391c5f94249bd7dc0475caf9c43f839a745cb210055959073d5ef602d12f1bf8db2c7e60093f5fbffa27ecd3575d6ca03dd468f854cec88606babd590edc6fa795f2525fad1dcb81cab9e980739e1622c9650d335f8f00751d87bfb980fd1940c3bc7b06bb73062faed4e73b635ebd5109ed8bc8c21c4ec776345cd488a59617224143305d9fdcb064e65b9109dc2389103987ee41468d1613987bffdd8bfe7929e7d2d799fb22a9bc8a400abb17c962497b5d6579c55db7979b8a60949ee31759f5013873d5f610056ab9057065381f86cb44c5c392a855eccb76d40276b0be45a6919163a5a2a60c7e735332c87cf8f56d01dc8122f01a55444e9d8857acff85d3776cb6973840969898b6e26cb3c2b4272481ac3f2ec0d4a431c37a6db5b543af703245b3d2e6f1a1f122cd4cdf978285aad0ef7be6e5e5738a41fe8a941ae84811dba1ba265351fc85d95325c40fc1d737992fca90fa0738b8cd2108f9238490ff4c7a84d56e6009568b6febaf03b723abf1742b8cf3311331b33c3b9cd1f25ccf25db2f06ec2d2b0664dffdf05831c51a1a08fc499f9e8e04c23b0f7b5106e65913a65baaf963e4d38b13c4de41cc39cacc0393909a61a6fb1d98455c4831b7f8295f95d1b5b6af46193fd97296fc66bb3319f9d1b75818ce9f3176de68d4f8c4f758b1157a4709222df95d3a34cbe094086cb78cad42b974b03828fe81afff1e8a733060d0f63a6a5d589cafd566b36c6eeb512dbee5805407a57859d2195c5e8032f961552e7585671490cddd0cc1e6f516dd9f838040fcaa94ce4367642ffcfe7a0c112a25381031465d990ab1a01b75a5966d04844065f1f118cf0d1c116b6c34f796957fc2228cfca4936d8036b0f4a5d39c12bd660c37aa8bd47a303583a520d3db871b2de6990098dbcaa66b5513e493b50da6ea8415414f95965a8424c727d48e1449780e8860513244b61429b8a2a6d0b881239e54882d05920e4730863f19b34f6f48d26e7c45a65625cd80ba359e5d47ea0373deced5f0a4e84f9718d2e6aa46adb3a9c87b650ffe925cd7063de5d72912c308ffa9f7d6641c9ec82e038f51493792b42fc8019b480d0d54985fc0f7f86c80aafefc4bb3f9da67fbe034d1713319ae95c47ced97e09d5e121212eb4f5e10a55987edc54f13739e534a9ab9839bf1f3a42f6311c748d465fc8e117a26766e73d59898e34d2eee291cd20cddd6f3ad1d07586eb4db733440a3179b53c87b39a185acfab259e70cf078f6b878ba8db4ae98d9748a13670feb16731117b9b10a957fd51394a3b412c836f6ce1af7070a2cb3d73b155f576d821d4491c8de854d02353f815974b6d5a560944d781530c1e7af0e302730ccb5aa38b4b1cc1cf452a0fc0e9a091798b2f4e35be94c8a8a72f07853ccbf3e226a0259a0051c7638c4a285e4d7f6b606e4eb707d51d8ef3b1df65e1cb48849a29de75adbb19429c2e022a05c751459cd0ff536b42e2186aea7a299ead87ee2a050859b05da1698c94be83dd74b4b466cd962557ad32865966732c352e527fa0b0407a6b0022766c64566a3ec4861937747f2e07fce824b2ca0d8d12e3538c8a3d25a87f8332a9cbc83f124bcc0b886f4cf82c0d6110cb8f6f0cf18aa3dc88f66a8e3f59fd4480c3adeb1492da7c7f177870cd9808f3231eef91ad6b622bcaf55643e3ddc8d3b1f8b57602eb4fffa127b78d8fe424c385e4e577f8f31d424c6f4fd86aa0667d98ac2a6de2f598a14a160f2495f69888d432812493a0ca775d3cdf40818c7ff9e5a2164f715c993625588a68ed971849330cfb0d32ae6360a3b7c2b0ea413242291706ebc4ad649ca1a57e9de7d99f574d3891bb2912c921b7a7f9a90bf7c428b6970b21b8158b0c913caeff7ae6dfc745d367cf5b0b93c4b39e8c9a0788360aa41785960326d4abb8c16a9dc25aa4f63609e5909966bede3a99084ac9b3040b68aa7ad5bd7cd79e932931bde3accd9a46db66b19c5919a68992452e1a4d3b1506ad99c4a56f66f8cdc02f50c3e904957e3fe3e72d0ac3f6f10c5dd838afafaa5a8a052ceb37fa97fdb9d8e71004aba543ecca85ec88d69db7d46bc1bf089a614a397ffed1c89f43e0bd3f5668d628f22574f713fa793d3ad8ba14c08bc7071424e4dcd821e5d914b29e6862afee6e2f695544e28de50dee2c9b31ac3d36b4a53188dd1f602ac3faac101218b92e19b396353680dba8b0d7a86d00222832f560269362a07c12b25b9a42d748dfc98b8cd05ec713a7f4e17fefb7ddd49a7a38a72bd11d8f2da98aff2521ed44a2dda6c5fe56f7eb78ce6f5945e2a0e04b7e74e692640c109560b8037a0b995bb50b951c765225a3736a03c5c999328a6e404d3627da9aea8e110ece0d57296a86c780c16e6d21c6f9d3c1e6997e00a9bcc40b7b02e3964e25f291dd2e658c8cdf3497d35c2597923fcf8a87223bde8166933754710b2cdcd0b9935f0c04cc97ce201966f0b945708875a428394b4c7faf661ba8b6d35f0ba5d4f6bd51502ae25893ba2b32147689448410a40a09e14b4077c11f271fcb46cddce4ae16e329c09061f9398916ffd6137e6d106afbe97db920abb260c1a218767449fbedb7919518c1bbca93a8b7f640468a0f987a8258d55ec2702ee7ad386f412b07d77920ca6cb2c36cb426de4f3e90092db6e8d2d786d520d0dc0441bc98dc6a10fec47f87b9ead5c11f232035225f826fab74b4a45279344cb38174ddb9fe7794017898ac714e0892098fa9f5f522d6c16c9af130f3bcd225d9f40324edcd7e7aa8031bc30460ba4285e8732fb5e3d0cb7736b05540e340ff7ed7fa540e166d1eb053f9c22e342e7d30d7bdc836ea5f1df25233e6e50a3520138a7eebe0ef6d72e8699b0f68b4fc99e899ba8d1669cbe7c9af296391083e15b5374162c03e1b0e5de3f0f250eb1d040cf2a4b8cfa4799c8281a4e8c92bc244730814d19eb2c14ecdaa525253e112d12757d49d92b4ae22b502781d5c20d5dfaf6f80d8db1a5b3dae00dd3146438e2cc573c93a1d7c3057a308255d002057f5f491463f7a018bb63ec9008b05a089b70512851c6449b4127efee6a8baaf4afc5581c49c48d5bb1db182bc936ddd0505ab7a941bd29b7e674cdf4c3327aaf53364271fd79f7f21fce359de940d9a82d17bd0df7cf25c27c45c49e3664178464ca791f809906de74bd710926561d5ffb651a37b6dbf12db36b7e3de67027b6a5557cac24cda44b86d741bbdb9ae65cca48d355077bea2d606f0e91357d47430fe7c7ee8df80a746fa4d553b089729dc9b3a4cc980314bc53e4a3c65578687408f1038ce7571630486a14f187308e291b3feb9799d4642dafb99192902b86e8ac8a487637a7288710073e0e4d0b1174b08f326f9c55debbf66468f4c1da90762b64b91b99717022a463bccc9b6bff95ae10cf46759bde811f6b22846b9acb152e978a82afb91ae44a2ca1094fe6a757adea40c7705e3d2e5549733da91f22f4e34ec59b3045cfda255e06da9b8e5eb17a95a43775d67a9add8e4b530656a2d6a5ed5bc8d78595d6d2688dcd09d51173a6670c7c71b6913dc92c08bb4e4de9522b338c581356c37a7ddaf509050b50b0e74e401c0b76fb0062407664cb076808aa59991593d453c1c7e090bac17b88eb6b5e6f0ae9dc35173be30fdaa7cecd8285a7235a1d8979e6a769143fe938bb4047167cd01055fd080f7d8d435fe7d3e03a53e7ca4598cb3b4c265352891b6b40872625f157ae43d703aaf9c4376d9f02c003f3488b90378fa1c4936829c09c70c6fa8f049883c909fadc7e0b492e2823ecde921c6050b7f36dea0047a9a9bbf837a414d46bcaf3fd0faa4422ef11f4054294d25e553203f0f9354756521d4805c14a7b694d0808c55455e06cba9d93ce10b3ba082f25ff35feca2e4413605f3e470715ad9f1c10d6feb246e848b4d9c138fe734640ff125308e157c1f22c2e49acd2ecce153d0ea9ef4bcfb83b8faab3e729c03d85f3791ace97ee6fb594bf8510576790d4c996c0179c7ed72ee90ffdb5e47dffa3e0baa9d4bcec2630c51ec8e7afeeda361eb73d5215a70ab010ef60a5b3b291adb4a183df513958ee46349e290da8ef362fc62baa4ac788dfd06e7b385f04a3f95222a38be90eba5db9fb58b674d09b5ba05c8b74459ae8d08234a561ca87665baeb401740fce053bd7574353c8662be0c48e67acff8ca07330825882801a4c7b845213f43da8913df086b48485046ea423fdfa55523a1930f283736a7a5614dabd3abd2901227bd8aa3331eee00a3fa3fda70e91114e89deee858dd5d199c566aa005af39fcb96def00c77ab79b5f3a48e2d679c8c2f00e115211188726e1d03d5173407f6166b468a4115599b613604008e46a909faf4b6587b55727c103f166a907928955988bc306ca654b250efb6351e73d5dd683a742a1749fc64080230f9892518471eac0600c1d6bee5dfce0898c503c455c5f460564edef7c2134cc389f19e1c4b05890458bbd9c9a769ee702e796e968bc3b6791b4c710fe9cefe24d596f3e4888584219bc28bbbeeaa97cc58b6c38f01531ded1fb781a73b82bbc4ac0fca9e005205bbee81197ea88171eacf6b65cefe01effeccf346c887dd900edbb39fbef8ca53e9cd734a7de98efe9893c72589971c7baa00ce100070fc6a905be104f8a168f8238ad6cd21eaa5f5906aa7000547b2aaa8c2f55f1c8dc81b25a87f16acd3d2711e392ae25c03dbab151c9a6b99e862c0f8da5851d903a3805c03368d10c74e0e650596c26c8829bf5b37418f2b7f7da24d85f19594d808a1eec88827653daf5a56a263c42cd9322d4f3fa37860d80f5a3ba0c4b9007f2355c181a383aba2b57b892fdb6f366f52f8066a5a092a9b123472044661be47970ddf3ad84e1b72b1f47b97d42420fec3f50cdc02f9199b83359510fc684f7e70485c6ddc9624b86a416411bde08f402ac8db8ad7ba15558921d62a59c968835c7ac40d72e74ecda85fa9ea15b9bb992339306b6727e559e19ad62e62d514f55fea9d5ca53b2547ac4842f14421e876aced3f1ec39e047541ec38b5e3595d4c272179e27eab4dadd8e0ab2f7789e410ec7c3f2392c478d6fd0373a1ac956ca218a6684fc9c37cbc644b7b3c0c3731e343477e0301d7ae5f512cfcea836c93f65d80595b92452ef7065c924a0dad1f5adbad200c2014b3843e807f77b3fc369af645f25673ab6e874345d47cc61e5461b537565fc84c9da6976bd404755deb569e696296d372ae7413de74deaaab241f375d32fdaebe1b73bd361e92e1d09838f0f4fe96011b403837994666e7ab5", @typed={0x8, 0x7b, 0x0, 0x0, @pid=0xffffffffffffffff}, @generic="cee42dd28e3f52aee8d7a91f41360c271dc994d6c09730284216ae66b25938a87480ed76f5bb06ade0"]}, @typed={0x15, 0x9f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}, 0x1424}], 0x5, &(0x7f0000003140)=[@rights={{0x30, 0x1, 0x1, [0xffffffffffffffff, r1, r2, r3, r4, r5, r6, r7]}}, @rights={{0x2c, 0x1, 0x1, [r8, r9, r10, 0xffffffffffffffff, r11, r12, r13]}}, @rights={{0x1c, 0x1, 0x1, [r14, r15, r16]}}, @rights={{0x18, 0x1, 0x1, [r17, r18]}}], 0x98, 0x400c880}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r19 = socket$nl_generic(0x10, 0x3, 0x10) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r19, 0x10e, 0x8, &(0x7f0000000040)=0x4, 0x4) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) 00:08:52 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.414927][ T137] netdevsim netdevsim0 netdevsim3 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:08:52 executing program 3: r0 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) r2 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) r3 = openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) r5 = syz_open_dev$ptys(0xc, 0x3, 0x0) r6 = openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2", 0xffffffffffffffff, 0xffffffffffffffff}) r8 = openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r9 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r9, 0x0, 0x0, 0x0, 0x0, 0x0) r10 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r10, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) r11 = bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) ioctl$SW_SYNC_IOC_CREATE_FENCE(0xffffffffffffffff, 0xc0285700, &(0x7f0000002f80)={0x10000, "deb61e2f2a84540b872481f48e71267ebf181545a2e58b1a73b1d5c64fa91704", 0xffffffffffffffff}) r13 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000003040)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x101, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x0, 0x3}, 0x48) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000002fc0)={{0x1, 0x1, 0x18, 0xffffffffffffffff, {0xee01, 0xffffffffffffffff}}, './file0\x00'}) r15 = openat$autofs(0xffffffffffffff9c, &(0x7f00000030c0), 0x2c040, 0x0) r16 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r16, &(0x7f00000046c0)={0x0, 0x0, &(0x7f0000004600)=[{0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}], 0x9}, 0x0) r17 = socket$nl_generic(0x10, 0x3, 0x10) r18 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000003100)='./binderfs/binder-control\x00', 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f0000003200)={&(0x7f0000000100)=@proc={0x10, 0x0, 0x25dfdbfe, 0x20000000}, 0xc, &(0x7f0000001800)=[{&(0x7f0000003240)=ANY=[@ANYBLOB="480100003c00020029bd7000fedbdf25c80059800800b60005000000fcef61d3255f79e749b74989cefd2d67f4a3a7f5d872130e436ace7f0251338b726457f2101bcb86d236c5477fb7149c43d955ff467b1d1d809fe6128edcf1b58e28e82300c5c2573e2dc0ad416e2f2871b5d459581f24b06e903ecd8649a80346f46972ab1e0e864198d13853cee89e3a595105e83061e0793077e3f466228c08009600710000000800af00", @ANYRES32=0x0, @ANYBLOB="0d00660080000000000000000500100000000000150098002f6465762f7669727475616c5f6e6369000000000c00c800b8000000000000000c0005000100000000000000c866ef508d669a67a13484453afb67d337f7230ab1bd797f79bcaa1764573b3e11364e29f3db8e20a98ed041912cdf40a5c23af50e2e3a7d2e9315821ce9e21b034756e42c2eb5d5f961a7cae21fbc08bed3e4ad7e2847c6834e35f695631a9b6f03070000"], 0x148}, {&(0x7f0000000140)=ANY=[@ANYBLOB="18ff20471b34cc000000200200022dc77000fedbdf250800e4065f3a788261d9ff0f971ca11525fbe80a9819385c78253494efe3e46d60cc422d35bb9907341cc1a20139fdeb7fbcb083617b8a49cc448c78ba24e08b35f47df84e534deb9226311e8b2aaef55c3a6d1987a75e03ffd476638d92a519543bcff12c98ad38da7835761dd0cdeba44f4d8e9c00000000e64daefac1c90189ce5bd8cca7dfc5689f5c6e6e3cc610249e972f824fade4f3920c46ebb50e3fc50791edff0fcc52b365e1da600d02eb7f000000"], 0x18}, {&(0x7f0000000400)={0x20c, 0x10, 0x100, 0x70bd25, 0x25dfdbfe, "", [@nested={0x40, 0x131, 0x0, 0x1, [@typed={0x8, 0xf2, 0x0, 0x0, @pid=0xffffffffffffffff}, @typed={0x8, 0x9a, 0x0, 0x0, @pid}, @typed={0x15, 0x27, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @typed={0x14, 0x58, 0x0, 0x0, @ipv6=@local}]}, @typed={0x4, 0x64}, @generic="c789e7cc7ca8dab63be9ed69e6e9a202e2d52f41aa", @generic="bfa355ce6fab563c093e78c05135747bfff952128f788a9c1ac249a7191e52772e6dabd2567244773291d880df05c45846fb4f3f987662064e2bb43b7ed60d3c4b6e54c4409c3ed0d40e512cbb060d95117797bd7f17b73f40f7dc49ea6c0be1efabf891a76d6f8550dd5d4686af8f9075a2c165e3a108c2c6f279672edfa2e65ce7a96a1ae680700aa7fe2a58ce", @nested={0x114, 0x105, 0x0, 0x1, [@generic="924193881425028ded5e45af38309971b787cba9b5eea3c02361f21ad6ec6bec8f6a34d2be156296c42513da5533fb9d3af4038f7b02c9fd412b3b70f94bb65e569922e3e9f346c0b8ed2ad109e0d5f3b953e5411ddfa7ed7b0a60542e24b0510f3d1a96629ba0a58021eb623f3acd60ad8f79c4278dd5c9213020d6359fb392fec783bed11fc3484972230b510f0e63cadf00b0b65fa135c61fe0cbf3280ed4a745719d8945f1fa1864a507ff66b293b5c51bdcb35a527746d528fb38c049829187d0a677b124fa68b2d6ceb565e1321b86fc", @generic="4d721c8b383769a764937f17db741ac512ad5f84dc682af8edeff4a74f6090876cc97ba3b128eecb2eef11038b50a328d818c4776ad01890bf08440a9a"]}]}, 0x20c}, {&(0x7f0000000740)={0x1084, 0x10, 0x800, 0x70bd25, 0x25dfdbfd, "", [@typed={0x14, 0x5d, 0x0, 0x0, @ipv6=@loopback}, @typed={0x24, 0x18, 0x0, 0x0, @binary="17089d9abf899777e6a2d77f1e7d5724872ee91444fa4da34531d08dc7f9ae85"}, @typed={0xc, 0x128, 0x0, 0x0, @u64=0x9}, @typed={0x8, 0xe9, 0x0, 0x0, @pid}, @typed={0x1004, 0x25, 0x0, 0x0, @binary="4d25059a1a01986681c822bc09f098bb73c119783f81dbe8824a8fa9014888c3f670af704119f6cc1545d35174c987fb52d9051bf3b01e4c7a424f595280a47cc3a01b45f06cba1defe1ab7ce1d9f66c00f7e039720b0dfc9e1d53cb304618a32f26256dccd3da1d40e44d66bf9f3c413ebe925e39394fb082f7818758d0b7e58df5069d078aaf74adea049747759263eea91b63f86a6ddf0b3290685b897c6073700c5c1dd8fea2dd1d01602e4299d5f125cf0163fa2c1ebc6de3f042eb9bde2e479ebd7a738f35c4cdb587bef8e571e0edaf3c4705553180e1cf9bf0adf27fd2835864e25fad08d8cfaac58e6c60d720aa7903db9ccb41fcf8793e14f97c81867739fb11c90d583f0a35654979a5a00e8e43e542077da3c7d6f7466d5a00d5d60a23b518e501fcd3e489c97762f6173cbb973c450b3b26edeb7a730407c74e7f11aa25de3bfa907086945e67b54a837c73567ae62633dafebad4d38ab5813bdd9d80262b6d2471a1a61cb97dc4fcdebf0cd980c04319b0463b1d0db04131c411e535438233a2028650ace7688bec3de1806e44ea79a63aade453acfcdbf5aa3e63c3e075a5caf2bc745d88554b7d23afd40094c9e654a253050ccc8e1d549c6042c9f8da7cbea3898c8573b04b30a678bf6d39c0f68fd2cded90da3dc4a008dcb455a30765aba41fdf7b9324d7ba34ac8dbbd47bdbd1b022b78ce2b765342dd776059d3a79d86b317e199147b1933b0b1def209de8492bd6962fbc93338eea9a5b3288161294459c765bb8a982f8b16f258aec9efbaaf651667758f05a7e11163dcd36c75bccc3a6df491499ec6b5c9e86dbb37a495033bcf5fa1a4ab5de98a8810e717c7f8db4a22076ddc32271f4fcf615c0a80655a7bb304f1d020587f34e6e7c1b99b6102ef51a8d92d7373d849e170129f4a1801d45d5aaff670c07afddfdc55fb0a597951c58603585f247f5a11f9c35ec83a1680f103f8ecf6deead4ad441cd67a4049486a16ccb6ccb5e935bf3a002312a2c419a0b14dd436730bd9b7f5c74c36f6b3dbcd184d750d0278fb1f8184d0e7545b7d7af48d71ca0372ea5b8e9d8ad797dafce2418bb24513a2653a7786998669795d8146f5b7ce28fc183aada6a942d1e0df98c1dedb579a46ed84a97798e6c4415266c114c659b01cb35c88a6c4b4b0383c5fc2b06c551319ebc9040a9d2939cc8bc1f1e5d0e156ef1d4d782ed7d87a28f3ef15a4494ef55ec26eb665078fa23548b9a66e438c3f747ff20ab2ea6aa03298b45fac67ff91dc3aee6abc127d5ddfd1a6f8d91999f85bc95483ddf64ee31f0a3d93312b5b29083233478ac85e37f9550f12cebeb072d24932145c7a211f4420f2bcabc52cb4caeb92354021181554ad0a4ae1c0d9d90707990362236d42cb29147f27d6830a4123d0ac346ee69a91922107d80b3457a66f274d98ef2d867022b34379d229ac4e702a64db3bf011bd1b4daf0a6d878ee9b681c9f52e5560a20225e871fa7ac880881f1e39a602400ecfa89665bb2558181232ded9fdf3a1063f4a1ff36241fd0ad58f09c05471dfc263a52dee36cf97f0915c22e7b22abfece46d5b0650fa1517a0fc3f820d5c7b378256daeee7723d9f04858747d630d3fbfcae75210663075676d890c492a50a0548f039deb20fe0a33017c1c721df39883dd6484267e05fb244370fcb729b522eee69ef47ec7934a380531a5f42b139822b6c7cff00d5cfb291ba9e2edcdf8b8b0735e49ccc2dabe9a1c363feeb821b04a645d3686544a212e887aba2bc4b6f171873c465a8a70e0810f036a73eb86ffc67182460373fb3a1b858eaac511934177126ac3060b11480fbc4474b558c496a27221a55d72370b498e081ed17d1197a42b373ccb86d76a2487dd66a60acc61bcbe7f61b72bd9e533793b7afc1425f1e8133c791cd65b4e766642c9497fc5ccca4a530d0235a620d2b08e2bc6e2b940d7e2b6312d489f63d525ea2e8c4df4d4ac05435b63681b91b13ae0d203eaa722662cdc33b8ac68fbaa2ff7f657952dd46650bb09633c519cc7c44126ee05babc55a3fbc2933bedf7cffcc55d7a3fcc4460e040e2896d1d8070246c821eef4e6723bc1cab07b305a4fb4ac6d78898d12144f56e25c8c5df9e6c80c7f866c3961ecbbaf1eed026da06b776c5591e1af5a7954bbb4efb0c1af303b688aff0576b1c50b79e6e310ba020ce3dcb564dd606d650a3dc2716d157bf013211468536ad0e6646213eb3fb02a884dac52b54a5aac329ec1eb4a5fb2beaa82f4a245410b3e583cd3c0014fca640d7a752ca6dd47c271d1d4ada9bb79b42501cb4f5894b9b9681a43bfae67b234288ee50e0101decda623569d5d9dad0a7e4c608858fb04de1cf356c9f0db9ea9cb5df287fb23efa0b17082c1257d77c16539bccfd3e7591540eb82e9f765dbb5bb289f2bd489a0488032b5461d1348d670693f4e53341f2456ace8e2f59ac470dc8c999e9e52c2e0e320a0e2d1238296d1185561002d93825b147935945c6d0139fe99d3a155a5c1afc2f3f176dd521fe4927737d83d33ce48d333f39158f5653f8345b7cda03ef8dd217868e7b8e06ca3b771fb923458fb649aeeca8e9f08ff2db3872ffffa6e7f773c37da792fd1c3030fe94f1ef0b971b4b6632783f15dcadfaab99be10db107957061340730f37c1c5655117dc0115f33987e051e6acffaaff9491694e4de23d08a24a986c1e866dd1910d215e22b591069971355cb145f52fc22caf8f5b120a9f4f9723cae79e1d15f05a2494d730e31e548148e9b326f37d9565157d03415b9f62c631d41c7d8d0b40b5b99f02dadc6ab95d81a97098628cf1c2803228169e4aad29f298753ef948a89c9dd38d4e953b0f3ec6d97a33d49c097baea530ad9f766d51f695976398d8e10f8c3fcf16b2bdf8f1d2535accf1fa0e85edbc4f189dc4ce16c9a7330fc77864b67a6b296b1349efc28d4281320214024fa851f467d3acb8d3b56ec98ffc73a330d8c395c5b2b264b24dcd6d03d4a7f558b30be5b22c6707b42e855a08df63825a89157bf76031f574454b142f05fe90bc9eeee2d7669615aff7233e6c01dd10da16a0fb5b488429e1fd936fc60b4897bc62c708e6df7d244e1631879bed247ae681bf434196d8db50ab1f702775c567bbc7f811fe4d00682f1d0084f67610318bee61189ac300da5cacacf85f067b5b21c6d68cdd3d11d48f365c7a187c2895765131800ae6d4b64c37ab8211f3d870284edbce98669e6721969e5690876f6807b9233784bc58a50cc75350f21622cb9e582644571add1159c307e0c2d385eef4b92faf94826b52fe7075ee40d891114e9ca69a389e1d571cd4c66bee43876b1251828a1a181b4011197c3fdd2ee610dd463320afc1e2f17c1f23ef6b4dfb1a73dc8747c94060a741cabedbad0114bfff358d0508c977b784f0d352de6b7c1e49dfe47e99211483be99e7b33147394f0fd9bc87f3cf0c02d174fbb09ff70add88ac6d5e243245b09d8e2b0d349698bd38f75ccfbd989ceb1bc4bb5732eddb177f97c80e0438d5e1ba15c9ff2ff50ba19ad93b4356a2664a73c8cb33663097887d7fe88c1de3d78309ebe7a16f483f544406ac448aa9356bd9171e4aa027e43295bc260bb1b6ba1fb00a4ac4490d1dd47e84bd6593231e1ed6e1e4d1defe56d7843e5a0120c2d33a9303349f424fc4476f27e2454edf30954496e5b4e789802eae2229f43073ad5c1ef64ffa3f8b4d0219b8812f2991866c36a7b1dfd8b510f69d1d1549d26db9d3c3cf7aefe1665c81082f2b9fe58d0e7670e9fdcda32566cb396a5ff81852b93945c6720554f36dcbc79bd44b5145d9e8382d6277afa4eb6a28d36f8239a533be4f9ee1fe8447ed63909391208b8b3af9df22e834ec9718f3474e73532fee8e113066d8eb5c317e97b56e7aa0749b4d84d8dd5435000c05e6b73ad4978cdbcc47480737923eafc754ea4587ce6b9b08269a3a73838e93c226d50449eeda62858617f98846c0658e9ffa8ecd2559b21fea3db08f413b593fa8c3e09d572993fe7acb83da43d75a5a789e5bca383f32966271ef7d4c1c6be53b98a1a7ec2c23957cccf019f7a7c0f9a87d89f1c15bcedbcef2bc4ce081aff354bc3df005ec813693348af7a4f6fa81ae4ec3616d63b07e1397cab03dc36654503b38dfc02250e00ce7e6d73ea53437b467df05979636a9d50e0b326049db5e801c4925ef646d85eb02b46d40a596b5c9b05cf31f441b97afde728d4c4c8062603f18db5b60eb327bc39dfe963601561728d4537c1da7a6120204b8b756d950c8779aec13c1891e0c1dc090857d3a5c15c074c2c86698dd9dbb43b8d789498c500058a3ac00b04ba5b4c648a7192751fb1d32d40786774ea04ad350ba349b7c0e37c39cd25d2e93d9ff06b0be3646fad982ff23f4f98fdd724396b641746ee2d8f786e7e81b16b59f4e5990ec945bc0294145e63295618e6f94ea3ab91e8bf28e65a395a27716ee8cd93d8095ee9f8472d879a383290ba00d74185bd6a34cd205eda9c4812f37eb1216cb6af28c87aec4771d7e2d9f52e525a3cc9abfc72663e5d791d4350b4393c7cb5e0572eb96de78389d794f80ebd9bc92b49445f6b68c52e6c5a6c68cfdef5b7f59626fb2a8026b906bb8c0e86f62a364c16ba1bb9ccc8f8288cc6cced7e1ecd974090dabc3a517634ce63d589b7256398037402969a485e92ff6fc6d9a572d3b9d423c1f76ac1de491792f373aaa2bc89dfb343f321b9949918504f515897f0723beadecb233327cb163116eabb36bb6543cb0cb3c2cb1145797afcfc6d626797b28f5f03bc2cd1d7f555283233d50dac1dc1f805782952e8a34c0951ae8a6691c31fa1eac2bed1c33c8ee5596002b3c687ebcee2d47475276d1cc1627866b904f5c3330fdc47061b7e104b4c59a5d019317865e71c347661f0aaa5035cfb216bcc3427fc0cfe06fc8522f04e7118584183e8722dfea9bb629a365d0b77e0e4689598a407765da16483309770b1186262f9625d507f93da36f2cda11c97c498a3cff11d6f351346f882063a277acf0f80bd452e90a44c161893b4aaade86750aa6c1ddd6df3254e1dd4e3a18054a369fc9e98ebd81d7ec35bfecda1a6ab66ec8c63331369649446d9d0da3ab9c288b83c0db15863c7aabcc68c4249b8cb7d12b46d9baeb2752ed26d71ffe48ad6392693153b13e7c8cc326c12b5603cfc9f2172269f547d2529fc70291bb38e3bb3279b9d765969bbb62b0b905251de86f0032ec26b502e907ca47756acb5bf49ac9014690d0a112d5bcf748a0fc1a6a65e9b69280ace21e9b048d4c519f0bdfa4639c5075e1cd724f3483056bca45f273ebc304ad2c26a651c31559accc7170c4f227a260fcd6127d5283dbca6241631d2752902223ad84cc715b219ccb2cf7ee1f1624fcc63739d828c543bd2ba6353f784efe7dc0e9c7837d5fdad8bc35849b18299eecb7936231181a87a3db380760c98f02389c37c0ed6758ee80d5497d0e1b2f3367efcd0225cccc3a09db7a8c8801ff8dcd86d92b53def8b3a61dc00de3f151628603238fad3b1c1cd38d98a0da9e30ab6e7228eb9970549ff06391adfa918f1257a827626b9c304d0687d0c163c3563867c317cacd371313ca74d0d05326546e9df2f22f9b1c21de253991321a13452a3b800c64e0f33c4d9b232058ba40ab92b2010464b4ca8fc10de7baf4860d80ce4b948ef95f6c684cf3c061461eaf5341c09c596eeea339e626c2f4f7db554283d590fa9b83"}, @nested={0x24, 0xdd, 0x0, 0x1, [@typed={0x8, 0xa2, 0x0, 0x0, @fd=r0}, @typed={0x15, 0x82, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}]}, 0x1084}, {&(0x7f00000019c0)={0x1424, 0x17, 0x2, 0x70bd29, 0x25dfdbfc, "", [@generic="133f5ce256e48e186b57eb92e3d67968193a2cf400d7cf3041790bce471bc257c3438d3e71a7d8393ef1b223f84de683314deeff92769a9fd838660088a17f6c2d29881a441cc1f341dc6ee6f4c5844598db7afd982706b468a39ff3ea4ad1631488fad356047ccc217c45ca7627ffc75733e22f495b0f256e7210a8d146509c7af6c38561393519a3b182b78ec1dd9bd25477a3cf587b27", @generic="e88ec576b23c16f99f3075597fe305d7cca55f465552bc35edd73b9e4d5ce2b8b486a37fc83eb8ac46f197789352d0b11938a7ba84027eff3fad4ac1222bb0d37e0471e0171b48d59270a259b2744ff58a8de6debd54f67c78bb47fbd8121844e292a0cb7e8a95c97e149a4d8f3ff033d6975eac8d84cbd3bf146fff16bc4690f3f3ac657ca1b672ece1672fbdb017c389dc634bf6bfb31190643d731343196c3facfa4d7450b8271a53706e4d46f5f873e479f3e1f3304d18c78039dec6ce750ec093f530b30b046597087ca6a4cb344ced7bdae4bbbcb05d3ef335cad3a84c072e397d382e1a4badfd816f22f0e243e0fcf79bbf3a6ce8f2", @generic="143fdd42b4d1a20607d75360dc83c395c8c3ce8252a2ce7008653d1b5e53fb041a8ccac0cb52b2df15a1e74b2e4c43a4ea411c43ffd236f7c089da121b3494d15fb915c4a8f58de93e5a92789bc38b542602a9d8af3e242d4a2e", @typed={0x14, 0x7, 0x0, 0x0, @ipv6=@ipv4={'\x00', '\xff\xff', @multicast1}}, @nested={0x120, 0xc6, 0x0, 0x1, [@typed={0x8, 0x3a, 0x0, 0x0, @ipv4=@initdev={0xac, 0x1e, 0x1, 0x0}}, @typed={0x8, 0x75, 0x0, 0x0, @u32=0x401}, @generic="b6895074bb7a3afe9630c6f99260b1f06c93bc2c0722540649440807d7a24b97f13ff6d21106215013a3fd5699a435b16158e53afb0f6ae0ff043d6639be1fb09a7c68", @generic="2749fb2104172d73bb85cc455515bcaf861c1add239821475b56e3cb827b64c9e8c2862d0fe2709bf0adc4377a55e79ee35151f9038a77fd6c62c16c9c743284915f107162cfc72596d29436f2527a36488416958a5c68125d51c5144522d271b74c6073546aae75103ccc5369d329bc92d13548f43eb3ff88fb916774b5d66961283b0fbbb72316c3552ef5507be8157459bd55be305195f50488b1a1", @typed={0x8, 0x105, 0x0, 0x0, @ipv4=@multicast2}, @generic="fd03adb6efa816bd7ffbfd7b56324c6291c0d47943cb464325b3037bcd446f2e310f70b6"]}, @nested={0x10da, 0xd, 0x0, 0x1, [@generic="845dde559616927f6d889da26095d1f40f2712fc6c56ebc6a455054ab7", @typed={0x8, 0x82, 0x0, 0x0, @u32=0x3f}, @typed={0x4, 0x50}, @typed={0x8, 0x57, 0x0, 0x0, @u32}, @typed={0x4, 0xa5}, @typed={0x15, 0x10b, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ed4141bc4f65989a9acee3dac370708ba95194311416f97e58f4a032c20ce1e20cdda03f81f0af7eed5353f09df66fda0010f3adb90b890776cd0a4e3436e5be3ca67c868e0ac0d2afe1832180a82f908afcbf870998621b", @generic="035ae798dea74a23cca6ba07b89ba8e528f72a786258952651b2346be481a4d62a955679adbf5cb2c0a16f7a01a314546c02d732e756dfee390855b0d51d78a4f4725565deafbcecbb65ab1fa666fef767d6019ed11646ccc00c84e249142e19a026cfaad65b4d2180b754ab21aa2d8c6c89eb9b2de191e769a05af2d8dc62f9bc80a546c4dd13b1f23f75253431fe2e92c09a577c7ea70dd8d9a348ecf6e7ccbea1c0c67ff07ba9edf44e3b3e50cc631d175fbb1480363d63c066920b56209b292fd82d1f11c66b7155ac3740daafd76cb43e70aacbcf63ce4e63604b6f9b178a6bf3dd0ebe5261ecb144838673d88e77eaa8686a4f9bb7eee95ed1e5698275ba14d32366d90e12551165017478e4774d497dba0995f4aa962010697dc9d7743991ca5334f52acb6f46f1e76a6c7d353d9f4b2de20ee797c17c599391c5f94249bd7dc0475caf9c43f839a745cb210055959073d5ef602d12f1bf8db2c7e60093f5fbffa27ecd3575d6ca03dd468f854cec88606babd590edc6fa795f2525fad1dcb81cab9e980739e1622c9650d335f8f00751d87bfb980fd1940c3bc7b06bb73062faed4e73b635ebd5109ed8bc8c21c4ec776345cd488a59617224143305d9fdcb064e65b9109dc2389103987ee41468d1613987bffdd8bfe7929e7d2d799fb22a9bc8a400abb17c962497b5d6579c55db7979b8a60949ee31759f5013873d5f610056ab9057065381f86cb44c5c392a855eccb76d40276b0be45a6919163a5a2a60c7e735332c87cf8f56d01dc8122f01a55444e9d8857acff85d3776cb6973840969898b6e26cb3c2b4272481ac3f2ec0d4a431c37a6db5b543af703245b3d2e6f1a1f122cd4cdf978285aad0ef7be6e5e5738a41fe8a941ae84811dba1ba265351fc85d95325c40fc1d737992fca90fa0738b8cd2108f9238490ff4c7a84d56e6009568b6febaf03b723abf1742b8cf3311331b33c3b9cd1f25ccf25db2f06ec2d2b0664dffdf05831c51a1a08fc499f9e8e04c23b0f7b5106e65913a65baaf963e4d38b13c4de41cc39cacc0393909a61a6fb1d98455c4831b7f8295f95d1b5b6af46193fd97296fc66bb3319f9d1b75818ce9f3176de68d4f8c4f758b1157a4709222df95d3a34cbe094086cb78cad42b974b03828fe81afff1e8a733060d0f63a6a5d589cafd566b36c6eeb512dbee5805407a57859d2195c5e8032f961552e7585671490cddd0cc1e6f516dd9f838040fcaa94ce4367642ffcfe7a0c112a25381031465d990ab1a01b75a5966d04844065f1f118cf0d1c116b6c34f796957fc2228cfca4936d8036b0f4a5d39c12bd660c37aa8bd47a303583a520d3db871b2de6990098dbcaa66b5513e493b50da6ea8415414f95965a8424c727d48e1449780e8860513244b61429b8a2a6d0b881239e54882d05920e4730863f19b34f6f48d26e7c45a65625cd80ba359e5d47ea0373deced5f0a4e84f9718d2e6aa46adb3a9c87b650ffe925cd7063de5d72912c308ffa9f7d6641c9ec82e038f51493792b42fc8019b480d0d54985fc0f7f86c80aafefc4bb3f9da67fbe034d1713319ae95c47ced97e09d5e121212eb4f5e10a55987edc54f13739e534a9ab9839bf1f3a42f6311c748d465fc8e117a26766e73d59898e34d2eee291cd20cddd6f3ad1d07586eb4db733440a3179b53c87b39a185acfab259e70cf078f6b878ba8db4ae98d9748a13670feb16731117b9b10a957fd51394a3b412c836f6ce1af7070a2cb3d73b155f576d821d4491c8de854d02353f815974b6d5a560944d781530c1e7af0e302730ccb5aa38b4b1cc1cf452a0fc0e9a091798b2f4e35be94c8a8a72f07853ccbf3e226a0259a0051c7638c4a285e4d7f6b606e4eb707d51d8ef3b1df65e1cb48849a29de75adbb19429c2e022a05c751459cd0ff536b42e2186aea7a299ead87ee2a050859b05da1698c94be83dd74b4b466cd962557ad32865966732c352e527fa0b0407a6b0022766c64566a3ec4861937747f2e07fce824b2ca0d8d12e3538c8a3d25a87f8332a9cbc83f124bcc0b886f4cf82c0d6110cb8f6f0cf18aa3dc88f66a8e3f59fd4480c3adeb1492da7c7f177870cd9808f3231eef91ad6b622bcaf55643e3ddc8d3b1f8b57602eb4fffa127b78d8fe424c385e4e577f8f31d424c6f4fd86aa0667d98ac2a6de2f598a14a160f2495f69888d432812493a0ca775d3cdf40818c7ff9e5a2164f715c993625588a68ed971849330cfb0d32ae6360a3b7c2b0ea413242291706ebc4ad649ca1a57e9de7d99f574d3891bb2912c921b7a7f9a90bf7c428b6970b21b8158b0c913caeff7ae6dfc745d367cf5b0b93c4b39e8c9a0788360aa41785960326d4abb8c16a9dc25aa4f63609e5909966bede3a99084ac9b3040b68aa7ad5bd7cd79e932931bde3accd9a46db66b19c5919a68992452e1a4d3b1506ad99c4a56f66f8cdc02f50c3e904957e3fe3e72d0ac3f6f10c5dd838afafaa5a8a052ceb37fa97fdb9d8e71004aba543ecca85ec88d69db7d46bc1bf089a614a397ffed1c89f43e0bd3f5668d628f22574f713fa793d3ad8ba14c08bc7071424e4dcd821e5d914b29e6862afee6e2f695544e28de50dee2c9b31ac3d36b4a53188dd1f602ac3faac101218b92e19b396353680dba8b0d7a86d00222832f560269362a07c12b25b9a42d748dfc98b8cd05ec713a7f4e17fefb7ddd49a7a38a72bd11d8f2da98aff2521ed44a2dda6c5fe56f7eb78ce6f5945e2a0e04b7e74e692640c109560b8037a0b995bb50b951c765225a3736a03c5c999328a6e404d3627da9aea8e110ece0d57296a86c780c16e6d21c6f9d3c1e6997e00a9bcc40b7b02e3964e25f291dd2e658c8cdf3497d35c2597923fcf8a87223bde8166933754710b2cdcd0b9935f0c04cc97ce201966f0b945708875a428394b4c7faf661ba8b6d35f0ba5d4f6bd51502ae25893ba2b32147689448410a40a09e14b4077c11f271fcb46cddce4ae16e329c09061f9398916ffd6137e6d106afbe97db920abb260c1a218767449fbedb7919518c1bbca93a8b7f640468a0f987a8258d55ec2702ee7ad386f412b07d77920ca6cb2c36cb426de4f3e90092db6e8d2d786d520d0dc0441bc98dc6a10fec47f87b9ead5c11f232035225f826fab74b4a45279344cb38174ddb9fe7794017898ac714e0892098fa9f5f522d6c16c9af130f3bcd225d9f40324edcd7e7aa8031bc30460ba4285e8732fb5e3d0cb7736b05540e340ff7ed7fa540e166d1eb053f9c22e342e7d30d7bdc836ea5f1df25233e6e50a3520138a7eebe0ef6d72e8699b0f68b4fc99e899ba8d1669cbe7c9af296391083e15b5374162c03e1b0e5de3f0f250eb1d040cf2a4b8cfa4799c8281a4e8c92bc244730814d19eb2c14ecdaa525253e112d12757d49d92b4ae22b502781d5c20d5dfaf6f80d8db1a5b3dae00dd3146438e2cc573c93a1d7c3057a308255d002057f5f491463f7a018bb63ec9008b05a089b70512851c6449b4127efee6a8baaf4afc5581c49c48d5bb1db182bc936ddd0505ab7a941bd29b7e674cdf4c3327aaf53364271fd79f7f21fce359de940d9a82d17bd0df7cf25c27c45c49e3664178464ca791f809906de74bd710926561d5ffb651a37b6dbf12db36b7e3de67027b6a5557cac24cda44b86d741bbdb9ae65cca48d355077bea2d606f0e91357d47430fe7c7ee8df80a746fa4d553b089729dc9b3a4cc980314bc53e4a3c65578687408f1038ce7571630486a14f187308e291b3feb9799d4642dafb99192902b86e8ac8a487637a7288710073e0e4d0b1174b08f326f9c55debbf66468f4c1da90762b64b91b99717022a463bccc9b6bff95ae10cf46759bde811f6b22846b9acb152e978a82afb91ae44a2ca1094fe6a757adea40c7705e3d2e5549733da91f22f4e34ec59b3045cfda255e06da9b8e5eb17a95a43775d67a9add8e4b530656a2d6a5ed5bc8d78595d6d2688dcd09d51173a6670c7c71b6913dc92c08bb4e4de9522b338c581356c37a7ddaf509050b50b0e74e401c0b76fb0062407664cb076808aa59991593d453c1c7e090bac17b88eb6b5e6f0ae9dc35173be30fdaa7cecd8285a7235a1d8979e6a769143fe938bb4047167cd01055fd080f7d8d435fe7d3e03a53e7ca4598cb3b4c265352891b6b40872625f157ae43d703aaf9c4376d9f02c003f3488b90378fa1c4936829c09c70c6fa8f049883c909fadc7e0b492e2823ecde921c6050b7f36dea0047a9a9bbf837a414d46bcaf3fd0faa4422ef11f4054294d25e553203f0f9354756521d4805c14a7b694d0808c55455e06cba9d93ce10b3ba082f25ff35feca2e4413605f3e470715ad9f1c10d6feb246e848b4d9c138fe734640ff125308e157c1f22c2e49acd2ecce153d0ea9ef4bcfb83b8faab3e729c03d85f3791ace97ee6fb594bf8510576790d4c996c0179c7ed72ee90ffdb5e47dffa3e0baa9d4bcec2630c51ec8e7afeeda361eb73d5215a70ab010ef60a5b3b291adb4a183df513958ee46349e290da8ef362fc62baa4ac788dfd06e7b385f04a3f95222a38be90eba5db9fb58b674d09b5ba05c8b74459ae8d08234a561ca87665baeb401740fce053bd7574353c8662be0c48e67acff8ca07330825882801a4c7b845213f43da8913df086b48485046ea423fdfa55523a1930f283736a7a5614dabd3abd2901227bd8aa3331eee00a3fa3fda70e91114e89deee858dd5d199c566aa005af39fcb96def00c77ab79b5f3a48e2d679c8c2f00e115211188726e1d03d5173407f6166b468a4115599b613604008e46a909faf4b6587b55727c103f166a907928955988bc306ca654b250efb6351e73d5dd683a742a1749fc64080230f9892518471eac0600c1d6bee5dfce0898c503c455c5f460564edef7c2134cc389f19e1c4b05890458bbd9c9a769ee702e796e968bc3b6791b4c710fe9cefe24d596f3e4888584219bc28bbbeeaa97cc58b6c38f01531ded1fb781a73b82bbc4ac0fca9e005205bbee81197ea88171eacf6b65cefe01effeccf346c887dd900edbb39fbef8ca53e9cd734a7de98efe9893c72589971c7baa00ce100070fc6a905be104f8a168f8238ad6cd21eaa5f5906aa7000547b2aaa8c2f55f1c8dc81b25a87f16acd3d2711e392ae25c03dbab151c9a6b99e862c0f8da5851d903a3805c03368d10c74e0e650596c26c8829bf5b37418f2b7f7da24d85f19594d808a1eec88827653daf5a56a263c42cd9322d4f3fa37860d80f5a3ba0c4b9007f2355c181a383aba2b57b892fdb6f366f52f8066a5a092a9b123472044661be47970ddf3ad84e1b72b1f47b97d42420fec3f50cdc02f9199b83359510fc684f7e70485c6ddc9624b86a416411bde08f402ac8db8ad7ba15558921d62a59c968835c7ac40d72e74ecda85fa9ea15b9bb992339306b6727e559e19ad62e62d514f55fea9d5ca53b2547ac4842f14421e876aced3f1ec39e047541ec38b5e3595d4c272179e27eab4dadd8e0ab2f7789e410ec7c3f2392c478d6fd0373a1ac956ca218a6684fc9c37cbc644b7b3c0c3731e343477e0301d7ae5f512cfcea836c93f65d80595b92452ef7065c924a0dad1f5adbad200c2014b3843e807f77b3fc369af645f25673ab6e874345d47cc61e5461b537565fc84c9da6976bd404755deb569e696296d372ae7413de74deaaab241f375d32fdaebe1b73bd361e92e1d09838f0f4fe96011b403837994666e7ab5", @typed={0x8, 0x7b, 0x0, 0x0, @pid=0xffffffffffffffff}, @generic="cee42dd28e3f52aee8d7a91f41360c271dc994d6c09730284216ae66b25938a87480ed76f5bb06ade0"]}, @typed={0x15, 0x9f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}, 0x1424}], 0x5, &(0x7f0000003140)=[@rights={{0x30, 0x1, 0x1, [0xffffffffffffffff, r1, r2, r3, r4, r5, r6, r7]}}, @rights={{0x2c, 0x1, 0x1, [r8, r9, r10, 0xffffffffffffffff, r11, r12, r13]}}, @rights={{0x1c, 0x1, 0x1, [r14, r15, r16]}}, @rights={{0x18, 0x1, 0x1, [r17, r18]}}], 0x98, 0x400c880}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r19 = socket$nl_generic(0x10, 0x3, 0x10) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r19, 0x10e, 0x8, &(0x7f0000000040)=0x4, 0x4) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) [ 532.455452][T30288] bridge0: port 1(bridge_slave_0) entered blocking state [ 532.466412][T30288] bridge0: port 1(bridge_slave_0) entered disabled state [ 532.468547][T30288] bridge_slave_0: entered allmulticast mode 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 2: r0 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) (async) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) (async) r2 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) r3 = openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) (async) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) r5 = syz_open_dev$ptys(0xc, 0x3, 0x0) r6 = openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) (async) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2", 0xffffffffffffffff, 0xffffffffffffffff}) (async) r8 = openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) (async) r9 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r9, 0x0, 0x0, 0x0, 0x0, 0x0) r10 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r10, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) (async) r11 = bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) ioctl$SW_SYNC_IOC_CREATE_FENCE(0xffffffffffffffff, 0xc0285700, &(0x7f0000002f80)={0x10000, "deb61e2f2a84540b872481f48e71267ebf181545a2e58b1a73b1d5c64fa91704", 0xffffffffffffffff}) r13 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000003040)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x101, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x0, 0x3}, 0x48) (async) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000002fc0)={{0x1, 0x1, 0x18, 0xffffffffffffffff, {0xee01, 0xffffffffffffffff}}, './file0\x00'}) (async) r15 = openat$autofs(0xffffffffffffff9c, &(0x7f00000030c0), 0x2c040, 0x0) r16 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r16, &(0x7f00000046c0)={0x0, 0x0, &(0x7f0000004600)=[{0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}], 0x9}, 0x0) (async) r17 = socket$nl_generic(0x10, 0x3, 0x10) (async) r18 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000003100)='./binderfs/binder-control\x00', 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f0000003200)={&(0x7f0000000100)=@proc={0x10, 0x0, 0x25dfdbfe, 0x20000000}, 0xc, &(0x7f0000001800)=[{&(0x7f0000003240)=ANY=[@ANYBLOB="480100003c00020029bd7000fedbdf25c80059800800b60005000000fcef61d3255f79e749b74989cefd2d67f4a3a7f5d872130e436ace7f0251338b726457f2101bcb86d236c5477fb7149c43d955ff467b1d1d809fe6128edcf1b58e28e82300c5c2573e2dc0ad416e2f2871b5d459581f24b06e903ecd8649a80346f46972ab1e0e864198d13853cee89e3a595105e83061e0793077e3f466228c08009600710000000800af00", @ANYRES32=0x0, @ANYBLOB="0d00660080000000000000000500100000000000150098002f6465762f7669727475616c5f6e6369000000000c00c800b8000000000000000c0005000100000000000000c866ef508d669a67a13484453afb67d337f7230ab1bd797f79bcaa1764573b3e11364e29f3db8e20a98ed041912cdf40a5c23af50e2e3a7d2e9315821ce9e21b034756e42c2eb5d5f961a7cae21fbc08bed3e4ad7e2847c6834e35f695631a9b6f03070000"], 0x148}, {&(0x7f0000000140)=ANY=[@ANYBLOB="18ff20471b34cc000000200200022dc77000fedbdf250800e4065f3a788261d9ff0f971ca11525fbe80a9819385c78253494efe3e46d60cc422d35bb9907341cc1a20139fdeb7fbcb083617b8a49cc448c78ba24e08b35f47df84e534deb9226311e8b2aaef55c3a6d1987a75e03ffd476638d92a519543bcff12c98ad38da7835761dd0cdeba44f4d8e9c00000000e64daefac1c90189ce5bd8cca7dfc5689f5c6e6e3cc610249e972f824fade4f3920c46ebb50e3fc50791edff0fcc52b365e1da600d02eb7f000000"], 0x18}, {&(0x7f0000000400)={0x20c, 0x10, 0x100, 0x70bd25, 0x25dfdbfe, "", [@nested={0x40, 0x131, 0x0, 0x1, [@typed={0x8, 0xf2, 0x0, 0x0, @pid=0xffffffffffffffff}, @typed={0x8, 0x9a, 0x0, 0x0, @pid}, @typed={0x15, 0x27, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @typed={0x14, 0x58, 0x0, 0x0, @ipv6=@local}]}, @typed={0x4, 0x64}, @generic="c789e7cc7ca8dab63be9ed69e6e9a202e2d52f41aa", @generic="bfa355ce6fab563c093e78c05135747bfff952128f788a9c1ac249a7191e52772e6dabd2567244773291d880df05c45846fb4f3f987662064e2bb43b7ed60d3c4b6e54c4409c3ed0d40e512cbb060d95117797bd7f17b73f40f7dc49ea6c0be1efabf891a76d6f8550dd5d4686af8f9075a2c165e3a108c2c6f279672edfa2e65ce7a96a1ae680700aa7fe2a58ce", @nested={0x114, 0x105, 0x0, 0x1, [@generic="924193881425028ded5e45af38309971b787cba9b5eea3c02361f21ad6ec6bec8f6a34d2be156296c42513da5533fb9d3af4038f7b02c9fd412b3b70f94bb65e569922e3e9f346c0b8ed2ad109e0d5f3b953e5411ddfa7ed7b0a60542e24b0510f3d1a96629ba0a58021eb623f3acd60ad8f79c4278dd5c9213020d6359fb392fec783bed11fc3484972230b510f0e63cadf00b0b65fa135c61fe0cbf3280ed4a745719d8945f1fa1864a507ff66b293b5c51bdcb35a527746d528fb38c049829187d0a677b124fa68b2d6ceb565e1321b86fc", @generic="4d721c8b383769a764937f17db741ac512ad5f84dc682af8edeff4a74f6090876cc97ba3b128eecb2eef11038b50a328d818c4776ad01890bf08440a9a"]}]}, 0x20c}, {&(0x7f0000000740)={0x1084, 0x10, 0x800, 0x70bd25, 0x25dfdbfd, "", [@typed={0x14, 0x5d, 0x0, 0x0, @ipv6=@loopback}, @typed={0x24, 0x18, 0x0, 0x0, @binary="17089d9abf899777e6a2d77f1e7d5724872ee91444fa4da34531d08dc7f9ae85"}, @typed={0xc, 0x128, 0x0, 0x0, @u64=0x9}, @typed={0x8, 0xe9, 0x0, 0x0, @pid}, @typed={0x1004, 0x25, 0x0, 0x0, @binary="4d25059a1a01986681c822bc09f098bb73c119783f81dbe8824a8fa9014888c3f670af704119f6cc1545d35174c987fb52d9051bf3b01e4c7a424f595280a47cc3a01b45f06cba1defe1ab7ce1d9f66c00f7e039720b0dfc9e1d53cb304618a32f26256dccd3da1d40e44d66bf9f3c413ebe925e39394fb082f7818758d0b7e58df5069d078aaf74adea049747759263eea91b63f86a6ddf0b3290685b897c6073700c5c1dd8fea2dd1d01602e4299d5f125cf0163fa2c1ebc6de3f042eb9bde2e479ebd7a738f35c4cdb587bef8e571e0edaf3c4705553180e1cf9bf0adf27fd2835864e25fad08d8cfaac58e6c60d720aa7903db9ccb41fcf8793e14f97c81867739fb11c90d583f0a35654979a5a00e8e43e542077da3c7d6f7466d5a00d5d60a23b518e501fcd3e489c97762f6173cbb973c450b3b26edeb7a730407c74e7f11aa25de3bfa907086945e67b54a837c73567ae62633dafebad4d38ab5813bdd9d80262b6d2471a1a61cb97dc4fcdebf0cd980c04319b0463b1d0db04131c411e535438233a2028650ace7688bec3de1806e44ea79a63aade453acfcdbf5aa3e63c3e075a5caf2bc745d88554b7d23afd40094c9e654a253050ccc8e1d549c6042c9f8da7cbea3898c8573b04b30a678bf6d39c0f68fd2cded90da3dc4a008dcb455a30765aba41fdf7b9324d7ba34ac8dbbd47bdbd1b022b78ce2b765342dd776059d3a79d86b317e199147b1933b0b1def209de8492bd6962fbc93338eea9a5b3288161294459c765bb8a982f8b16f258aec9efbaaf651667758f05a7e11163dcd36c75bccc3a6df491499ec6b5c9e86dbb37a495033bcf5fa1a4ab5de98a8810e717c7f8db4a22076ddc32271f4fcf615c0a80655a7bb304f1d020587f34e6e7c1b99b6102ef51a8d92d7373d849e170129f4a1801d45d5aaff670c07afddfdc55fb0a597951c58603585f247f5a11f9c35ec83a1680f103f8ecf6deead4ad441cd67a4049486a16ccb6ccb5e935bf3a002312a2c419a0b14dd436730bd9b7f5c74c36f6b3dbcd184d750d0278fb1f8184d0e7545b7d7af48d71ca0372ea5b8e9d8ad797dafce2418bb24513a2653a7786998669795d8146f5b7ce28fc183aada6a942d1e0df98c1dedb579a46ed84a97798e6c4415266c114c659b01cb35c88a6c4b4b0383c5fc2b06c551319ebc9040a9d2939cc8bc1f1e5d0e156ef1d4d782ed7d87a28f3ef15a4494ef55ec26eb665078fa23548b9a66e438c3f747ff20ab2ea6aa03298b45fac67ff91dc3aee6abc127d5ddfd1a6f8d91999f85bc95483ddf64ee31f0a3d93312b5b29083233478ac85e37f9550f12cebeb072d24932145c7a211f4420f2bcabc52cb4caeb92354021181554ad0a4ae1c0d9d90707990362236d42cb29147f27d6830a4123d0ac346ee69a91922107d80b3457a66f274d98ef2d867022b34379d229ac4e702a64db3bf011bd1b4daf0a6d878ee9b681c9f52e5560a20225e871fa7ac880881f1e39a602400ecfa89665bb2558181232ded9fdf3a1063f4a1ff36241fd0ad58f09c05471dfc263a52dee36cf97f0915c22e7b22abfece46d5b0650fa1517a0fc3f820d5c7b378256daeee7723d9f04858747d630d3fbfcae75210663075676d890c492a50a0548f039deb20fe0a33017c1c721df39883dd6484267e05fb244370fcb729b522eee69ef47ec7934a380531a5f42b139822b6c7cff00d5cfb291ba9e2edcdf8b8b0735e49ccc2dabe9a1c363feeb821b04a645d3686544a212e887aba2bc4b6f171873c465a8a70e0810f036a73eb86ffc67182460373fb3a1b858eaac511934177126ac3060b11480fbc4474b558c496a27221a55d72370b498e081ed17d1197a42b373ccb86d76a2487dd66a60acc61bcbe7f61b72bd9e533793b7afc1425f1e8133c791cd65b4e766642c9497fc5ccca4a530d0235a620d2b08e2bc6e2b940d7e2b6312d489f63d525ea2e8c4df4d4ac05435b63681b91b13ae0d203eaa722662cdc33b8ac68fbaa2ff7f657952dd46650bb09633c519cc7c44126ee05babc55a3fbc2933bedf7cffcc55d7a3fcc4460e040e2896d1d8070246c821eef4e6723bc1cab07b305a4fb4ac6d78898d12144f56e25c8c5df9e6c80c7f866c3961ecbbaf1eed026da06b776c5591e1af5a7954bbb4efb0c1af303b688aff0576b1c50b79e6e310ba020ce3dcb564dd606d650a3dc2716d157bf013211468536ad0e6646213eb3fb02a884dac52b54a5aac329ec1eb4a5fb2beaa82f4a245410b3e583cd3c0014fca640d7a752ca6dd47c271d1d4ada9bb79b42501cb4f5894b9b9681a43bfae67b234288ee50e0101decda623569d5d9dad0a7e4c608858fb04de1cf356c9f0db9ea9cb5df287fb23efa0b17082c1257d77c16539bccfd3e7591540eb82e9f765dbb5bb289f2bd489a0488032b5461d1348d670693f4e53341f2456ace8e2f59ac470dc8c999e9e52c2e0e320a0e2d1238296d1185561002d93825b147935945c6d0139fe99d3a155a5c1afc2f3f176dd521fe4927737d83d33ce48d333f39158f5653f8345b7cda03ef8dd217868e7b8e06ca3b771fb923458fb649aeeca8e9f08ff2db3872ffffa6e7f773c37da792fd1c3030fe94f1ef0b971b4b6632783f15dcadfaab99be10db107957061340730f37c1c5655117dc0115f33987e051e6acffaaff9491694e4de23d08a24a986c1e866dd1910d215e22b591069971355cb145f52fc22caf8f5b120a9f4f9723cae79e1d15f05a2494d730e31e548148e9b326f37d9565157d03415b9f62c631d41c7d8d0b40b5b99f02dadc6ab95d81a97098628cf1c2803228169e4aad29f298753ef948a89c9dd38d4e953b0f3ec6d97a33d49c097baea530ad9f766d51f695976398d8e10f8c3fcf16b2bdf8f1d2535accf1fa0e85edbc4f189dc4ce16c9a7330fc77864b67a6b296b1349efc28d4281320214024fa851f467d3acb8d3b56ec98ffc73a330d8c395c5b2b264b24dcd6d03d4a7f558b30be5b22c6707b42e855a08df63825a89157bf76031f574454b142f05fe90bc9eeee2d7669615aff7233e6c01dd10da16a0fb5b488429e1fd936fc60b4897bc62c708e6df7d244e1631879bed247ae681bf434196d8db50ab1f702775c567bbc7f811fe4d00682f1d0084f67610318bee61189ac300da5cacacf85f067b5b21c6d68cdd3d11d48f365c7a187c2895765131800ae6d4b64c37ab8211f3d870284edbce98669e6721969e5690876f6807b9233784bc58a50cc75350f21622cb9e582644571add1159c307e0c2d385eef4b92faf94826b52fe7075ee40d891114e9ca69a389e1d571cd4c66bee43876b1251828a1a181b4011197c3fdd2ee610dd463320afc1e2f17c1f23ef6b4dfb1a73dc8747c94060a741cabedbad0114bfff358d0508c977b784f0d352de6b7c1e49dfe47e99211483be99e7b33147394f0fd9bc87f3cf0c02d174fbb09ff70add88ac6d5e243245b09d8e2b0d349698bd38f75ccfbd989ceb1bc4bb5732eddb177f97c80e0438d5e1ba15c9ff2ff50ba19ad93b4356a2664a73c8cb33663097887d7fe88c1de3d78309ebe7a16f483f544406ac448aa9356bd9171e4aa027e43295bc260bb1b6ba1fb00a4ac4490d1dd47e84bd6593231e1ed6e1e4d1defe56d7843e5a0120c2d33a9303349f424fc4476f27e2454edf30954496e5b4e789802eae2229f43073ad5c1ef64ffa3f8b4d0219b8812f2991866c36a7b1dfd8b510f69d1d1549d26db9d3c3cf7aefe1665c81082f2b9fe58d0e7670e9fdcda32566cb396a5ff81852b93945c6720554f36dcbc79bd44b5145d9e8382d6277afa4eb6a28d36f8239a533be4f9ee1fe8447ed63909391208b8b3af9df22e834ec9718f3474e73532fee8e113066d8eb5c317e97b56e7aa0749b4d84d8dd5435000c05e6b73ad4978cdbcc47480737923eafc754ea4587ce6b9b08269a3a73838e93c226d50449eeda62858617f98846c0658e9ffa8ecd2559b21fea3db08f413b593fa8c3e09d572993fe7acb83da43d75a5a789e5bca383f32966271ef7d4c1c6be53b98a1a7ec2c23957cccf019f7a7c0f9a87d89f1c15bcedbcef2bc4ce081aff354bc3df005ec813693348af7a4f6fa81ae4ec3616d63b07e1397cab03dc36654503b38dfc02250e00ce7e6d73ea53437b467df05979636a9d50e0b326049db5e801c4925ef646d85eb02b46d40a596b5c9b05cf31f441b97afde728d4c4c8062603f18db5b60eb327bc39dfe963601561728d4537c1da7a6120204b8b756d950c8779aec13c1891e0c1dc090857d3a5c15c074c2c86698dd9dbb43b8d789498c500058a3ac00b04ba5b4c648a7192751fb1d32d40786774ea04ad350ba349b7c0e37c39cd25d2e93d9ff06b0be3646fad982ff23f4f98fdd724396b641746ee2d8f786e7e81b16b59f4e5990ec945bc0294145e63295618e6f94ea3ab91e8bf28e65a395a27716ee8cd93d8095ee9f8472d879a383290ba00d74185bd6a34cd205eda9c4812f37eb1216cb6af28c87aec4771d7e2d9f52e525a3cc9abfc72663e5d791d4350b4393c7cb5e0572eb96de78389d794f80ebd9bc92b49445f6b68c52e6c5a6c68cfdef5b7f59626fb2a8026b906bb8c0e86f62a364c16ba1bb9ccc8f8288cc6cced7e1ecd974090dabc3a517634ce63d589b7256398037402969a485e92ff6fc6d9a572d3b9d423c1f76ac1de491792f373aaa2bc89dfb343f321b9949918504f515897f0723beadecb233327cb163116eabb36bb6543cb0cb3c2cb1145797afcfc6d626797b28f5f03bc2cd1d7f555283233d50dac1dc1f805782952e8a34c0951ae8a6691c31fa1eac2bed1c33c8ee5596002b3c687ebcee2d47475276d1cc1627866b904f5c3330fdc47061b7e104b4c59a5d019317865e71c347661f0aaa5035cfb216bcc3427fc0cfe06fc8522f04e7118584183e8722dfea9bb629a365d0b77e0e4689598a407765da16483309770b1186262f9625d507f93da36f2cda11c97c498a3cff11d6f351346f882063a277acf0f80bd452e90a44c161893b4aaade86750aa6c1ddd6df3254e1dd4e3a18054a369fc9e98ebd81d7ec35bfecda1a6ab66ec8c63331369649446d9d0da3ab9c288b83c0db15863c7aabcc68c4249b8cb7d12b46d9baeb2752ed26d71ffe48ad6392693153b13e7c8cc326c12b5603cfc9f2172269f547d2529fc70291bb38e3bb3279b9d765969bbb62b0b905251de86f0032ec26b502e907ca47756acb5bf49ac9014690d0a112d5bcf748a0fc1a6a65e9b69280ace21e9b048d4c519f0bdfa4639c5075e1cd724f3483056bca45f273ebc304ad2c26a651c31559accc7170c4f227a260fcd6127d5283dbca6241631d2752902223ad84cc715b219ccb2cf7ee1f1624fcc63739d828c543bd2ba6353f784efe7dc0e9c7837d5fdad8bc35849b18299eecb7936231181a87a3db380760c98f02389c37c0ed6758ee80d5497d0e1b2f3367efcd0225cccc3a09db7a8c8801ff8dcd86d92b53def8b3a61dc00de3f151628603238fad3b1c1cd38d98a0da9e30ab6e7228eb9970549ff06391adfa918f1257a827626b9c304d0687d0c163c3563867c317cacd371313ca74d0d05326546e9df2f22f9b1c21de253991321a13452a3b800c64e0f33c4d9b232058ba40ab92b2010464b4ca8fc10de7baf4860d80ce4b948ef95f6c684cf3c061461eaf5341c09c596eeea339e626c2f4f7db554283d590fa9b83"}, @nested={0x24, 0xdd, 0x0, 0x1, [@typed={0x8, 0xa2, 0x0, 0x0, @fd=r0}, @typed={0x15, 0x82, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}]}, 0x1084}, {&(0x7f00000019c0)={0x1424, 0x17, 0x2, 0x70bd29, 0x25dfdbfc, "", [@generic="133f5ce256e48e186b57eb92e3d67968193a2cf400d7cf3041790bce471bc257c3438d3e71a7d8393ef1b223f84de683314deeff92769a9fd838660088a17f6c2d29881a441cc1f341dc6ee6f4c5844598db7afd982706b468a39ff3ea4ad1631488fad356047ccc217c45ca7627ffc75733e22f495b0f256e7210a8d146509c7af6c38561393519a3b182b78ec1dd9bd25477a3cf587b27", @generic="e88ec576b23c16f99f3075597fe305d7cca55f465552bc35edd73b9e4d5ce2b8b486a37fc83eb8ac46f197789352d0b11938a7ba84027eff3fad4ac1222bb0d37e0471e0171b48d59270a259b2744ff58a8de6debd54f67c78bb47fbd8121844e292a0cb7e8a95c97e149a4d8f3ff033d6975eac8d84cbd3bf146fff16bc4690f3f3ac657ca1b672ece1672fbdb017c389dc634bf6bfb31190643d731343196c3facfa4d7450b8271a53706e4d46f5f873e479f3e1f3304d18c78039dec6ce750ec093f530b30b046597087ca6a4cb344ced7bdae4bbbcb05d3ef335cad3a84c072e397d382e1a4badfd816f22f0e243e0fcf79bbf3a6ce8f2", @generic="143fdd42b4d1a20607d75360dc83c395c8c3ce8252a2ce7008653d1b5e53fb041a8ccac0cb52b2df15a1e74b2e4c43a4ea411c43ffd236f7c089da121b3494d15fb915c4a8f58de93e5a92789bc38b542602a9d8af3e242d4a2e", @typed={0x14, 0x7, 0x0, 0x0, @ipv6=@ipv4={'\x00', '\xff\xff', @multicast1}}, @nested={0x120, 0xc6, 0x0, 0x1, [@typed={0x8, 0x3a, 0x0, 0x0, @ipv4=@initdev={0xac, 0x1e, 0x1, 0x0}}, @typed={0x8, 0x75, 0x0, 0x0, @u32=0x401}, @generic="b6895074bb7a3afe9630c6f99260b1f06c93bc2c0722540649440807d7a24b97f13ff6d21106215013a3fd5699a435b16158e53afb0f6ae0ff043d6639be1fb09a7c68", @generic="2749fb2104172d73bb85cc455515bcaf861c1add239821475b56e3cb827b64c9e8c2862d0fe2709bf0adc4377a55e79ee35151f9038a77fd6c62c16c9c743284915f107162cfc72596d29436f2527a36488416958a5c68125d51c5144522d271b74c6073546aae75103ccc5369d329bc92d13548f43eb3ff88fb916774b5d66961283b0fbbb72316c3552ef5507be8157459bd55be305195f50488b1a1", @typed={0x8, 0x105, 0x0, 0x0, @ipv4=@multicast2}, @generic="fd03adb6efa816bd7ffbfd7b56324c6291c0d47943cb464325b3037bcd446f2e310f70b6"]}, @nested={0x10da, 0xd, 0x0, 0x1, [@generic="845dde559616927f6d889da26095d1f40f2712fc6c56ebc6a455054ab7", @typed={0x8, 0x82, 0x0, 0x0, @u32=0x3f}, @typed={0x4, 0x50}, @typed={0x8, 0x57, 0x0, 0x0, @u32}, @typed={0x4, 0xa5}, @typed={0x15, 0x10b, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ed4141bc4f65989a9acee3dac370708ba95194311416f97e58f4a032c20ce1e20cdda03f81f0af7eed5353f09df66fda0010f3adb90b890776cd0a4e3436e5be3ca67c868e0ac0d2afe1832180a82f908afcbf870998621b", @generic="035ae798dea74a23cca6ba07b89ba8e528f72a786258952651b2346be481a4d62a955679adbf5cb2c0a16f7a01a314546c02d732e756dfee390855b0d51d78a4f4725565deafbcecbb65ab1fa666fef767d6019ed11646ccc00c84e249142e19a026cfaad65b4d2180b754ab21aa2d8c6c89eb9b2de191e769a05af2d8dc62f9bc80a546c4dd13b1f23f75253431fe2e92c09a577c7ea70dd8d9a348ecf6e7ccbea1c0c67ff07ba9edf44e3b3e50cc631d175fbb1480363d63c066920b56209b292fd82d1f11c66b7155ac3740daafd76cb43e70aacbcf63ce4e63604b6f9b178a6bf3dd0ebe5261ecb144838673d88e77eaa8686a4f9bb7eee95ed1e5698275ba14d32366d90e12551165017478e4774d497dba0995f4aa962010697dc9d7743991ca5334f52acb6f46f1e76a6c7d353d9f4b2de20ee797c17c599391c5f94249bd7dc0475caf9c43f839a745cb210055959073d5ef602d12f1bf8db2c7e60093f5fbffa27ecd3575d6ca03dd468f854cec88606babd590edc6fa795f2525fad1dcb81cab9e980739e1622c9650d335f8f00751d87bfb980fd1940c3bc7b06bb73062faed4e73b635ebd5109ed8bc8c21c4ec776345cd488a59617224143305d9fdcb064e65b9109dc2389103987ee41468d1613987bffdd8bfe7929e7d2d799fb22a9bc8a400abb17c962497b5d6579c55db7979b8a60949ee31759f5013873d5f610056ab9057065381f86cb44c5c392a855eccb76d40276b0be45a6919163a5a2a60c7e735332c87cf8f56d01dc8122f01a55444e9d8857acff85d3776cb6973840969898b6e26cb3c2b4272481ac3f2ec0d4a431c37a6db5b543af703245b3d2e6f1a1f122cd4cdf978285aad0ef7be6e5e5738a41fe8a941ae84811dba1ba265351fc85d95325c40fc1d737992fca90fa0738b8cd2108f9238490ff4c7a84d56e6009568b6febaf03b723abf1742b8cf3311331b33c3b9cd1f25ccf25db2f06ec2d2b0664dffdf05831c51a1a08fc499f9e8e04c23b0f7b5106e65913a65baaf963e4d38b13c4de41cc39cacc0393909a61a6fb1d98455c4831b7f8295f95d1b5b6af46193fd97296fc66bb3319f9d1b75818ce9f3176de68d4f8c4f758b1157a4709222df95d3a34cbe094086cb78cad42b974b03828fe81afff1e8a733060d0f63a6a5d589cafd566b36c6eeb512dbee5805407a57859d2195c5e8032f961552e7585671490cddd0cc1e6f516dd9f838040fcaa94ce4367642ffcfe7a0c112a25381031465d990ab1a01b75a5966d04844065f1f118cf0d1c116b6c34f796957fc2228cfca4936d8036b0f4a5d39c12bd660c37aa8bd47a303583a520d3db871b2de6990098dbcaa66b5513e493b50da6ea8415414f95965a8424c727d48e1449780e8860513244b61429b8a2a6d0b881239e54882d05920e4730863f19b34f6f48d26e7c45a65625cd80ba359e5d47ea0373deced5f0a4e84f9718d2e6aa46adb3a9c87b650ffe925cd7063de5d72912c308ffa9f7d6641c9ec82e038f51493792b42fc8019b480d0d54985fc0f7f86c80aafefc4bb3f9da67fbe034d1713319ae95c47ced97e09d5e121212eb4f5e10a55987edc54f13739e534a9ab9839bf1f3a42f6311c748d465fc8e117a26766e73d59898e34d2eee291cd20cddd6f3ad1d07586eb4db733440a3179b53c87b39a185acfab259e70cf078f6b878ba8db4ae98d9748a13670feb16731117b9b10a957fd51394a3b412c836f6ce1af7070a2cb3d73b155f576d821d4491c8de854d02353f815974b6d5a560944d781530c1e7af0e302730ccb5aa38b4b1cc1cf452a0fc0e9a091798b2f4e35be94c8a8a72f07853ccbf3e226a0259a0051c7638c4a285e4d7f6b606e4eb707d51d8ef3b1df65e1cb48849a29de75adbb19429c2e022a05c751459cd0ff536b42e2186aea7a299ead87ee2a050859b05da1698c94be83dd74b4b466cd962557ad32865966732c352e527fa0b0407a6b0022766c64566a3ec4861937747f2e07fce824b2ca0d8d12e3538c8a3d25a87f8332a9cbc83f124bcc0b886f4cf82c0d6110cb8f6f0cf18aa3dc88f66a8e3f59fd4480c3adeb1492da7c7f177870cd9808f3231eef91ad6b622bcaf55643e3ddc8d3b1f8b57602eb4fffa127b78d8fe424c385e4e577f8f31d424c6f4fd86aa0667d98ac2a6de2f598a14a160f2495f69888d432812493a0ca775d3cdf40818c7ff9e5a2164f715c993625588a68ed971849330cfb0d32ae6360a3b7c2b0ea413242291706ebc4ad649ca1a57e9de7d99f574d3891bb2912c921b7a7f9a90bf7c428b6970b21b8158b0c913caeff7ae6dfc745d367cf5b0b93c4b39e8c9a0788360aa41785960326d4abb8c16a9dc25aa4f63609e5909966bede3a99084ac9b3040b68aa7ad5bd7cd79e932931bde3accd9a46db66b19c5919a68992452e1a4d3b1506ad99c4a56f66f8cdc02f50c3e904957e3fe3e72d0ac3f6f10c5dd838afafaa5a8a052ceb37fa97fdb9d8e71004aba543ecca85ec88d69db7d46bc1bf089a614a397ffed1c89f43e0bd3f5668d628f22574f713fa793d3ad8ba14c08bc7071424e4dcd821e5d914b29e6862afee6e2f695544e28de50dee2c9b31ac3d36b4a53188dd1f602ac3faac101218b92e19b396353680dba8b0d7a86d00222832f560269362a07c12b25b9a42d748dfc98b8cd05ec713a7f4e17fefb7ddd49a7a38a72bd11d8f2da98aff2521ed44a2dda6c5fe56f7eb78ce6f5945e2a0e04b7e74e692640c109560b8037a0b995bb50b951c765225a3736a03c5c999328a6e404d3627da9aea8e110ece0d57296a86c780c16e6d21c6f9d3c1e6997e00a9bcc40b7b02e3964e25f291dd2e658c8cdf3497d35c2597923fcf8a87223bde8166933754710b2cdcd0b9935f0c04cc97ce201966f0b945708875a428394b4c7faf661ba8b6d35f0ba5d4f6bd51502ae25893ba2b32147689448410a40a09e14b4077c11f271fcb46cddce4ae16e329c09061f9398916ffd6137e6d106afbe97db920abb260c1a218767449fbedb7919518c1bbca93a8b7f640468a0f987a8258d55ec2702ee7ad386f412b07d77920ca6cb2c36cb426de4f3e90092db6e8d2d786d520d0dc0441bc98dc6a10fec47f87b9ead5c11f232035225f826fab74b4a45279344cb38174ddb9fe7794017898ac714e0892098fa9f5f522d6c16c9af130f3bcd225d9f40324edcd7e7aa8031bc30460ba4285e8732fb5e3d0cb7736b05540e340ff7ed7fa540e166d1eb053f9c22e342e7d30d7bdc836ea5f1df25233e6e50a3520138a7eebe0ef6d72e8699b0f68b4fc99e899ba8d1669cbe7c9af296391083e15b5374162c03e1b0e5de3f0f250eb1d040cf2a4b8cfa4799c8281a4e8c92bc244730814d19eb2c14ecdaa525253e112d12757d49d92b4ae22b502781d5c20d5dfaf6f80d8db1a5b3dae00dd3146438e2cc573c93a1d7c3057a308255d002057f5f491463f7a018bb63ec9008b05a089b70512851c6449b4127efee6a8baaf4afc5581c49c48d5bb1db182bc936ddd0505ab7a941bd29b7e674cdf4c3327aaf53364271fd79f7f21fce359de940d9a82d17bd0df7cf25c27c45c49e3664178464ca791f809906de74bd710926561d5ffb651a37b6dbf12db36b7e3de67027b6a5557cac24cda44b86d741bbdb9ae65cca48d355077bea2d606f0e91357d47430fe7c7ee8df80a746fa4d553b089729dc9b3a4cc980314bc53e4a3c65578687408f1038ce7571630486a14f187308e291b3feb9799d4642dafb99192902b86e8ac8a487637a7288710073e0e4d0b1174b08f326f9c55debbf66468f4c1da90762b64b91b99717022a463bccc9b6bff95ae10cf46759bde811f6b22846b9acb152e978a82afb91ae44a2ca1094fe6a757adea40c7705e3d2e5549733da91f22f4e34ec59b3045cfda255e06da9b8e5eb17a95a43775d67a9add8e4b530656a2d6a5ed5bc8d78595d6d2688dcd09d51173a6670c7c71b6913dc92c08bb4e4de9522b338c581356c37a7ddaf509050b50b0e74e401c0b76fb0062407664cb076808aa59991593d453c1c7e090bac17b88eb6b5e6f0ae9dc35173be30fdaa7cecd8285a7235a1d8979e6a769143fe938bb4047167cd01055fd080f7d8d435fe7d3e03a53e7ca4598cb3b4c265352891b6b40872625f157ae43d703aaf9c4376d9f02c003f3488b90378fa1c4936829c09c70c6fa8f049883c909fadc7e0b492e2823ecde921c6050b7f36dea0047a9a9bbf837a414d46bcaf3fd0faa4422ef11f4054294d25e553203f0f9354756521d4805c14a7b694d0808c55455e06cba9d93ce10b3ba082f25ff35feca2e4413605f3e470715ad9f1c10d6feb246e848b4d9c138fe734640ff125308e157c1f22c2e49acd2ecce153d0ea9ef4bcfb83b8faab3e729c03d85f3791ace97ee6fb594bf8510576790d4c996c0179c7ed72ee90ffdb5e47dffa3e0baa9d4bcec2630c51ec8e7afeeda361eb73d5215a70ab010ef60a5b3b291adb4a183df513958ee46349e290da8ef362fc62baa4ac788dfd06e7b385f04a3f95222a38be90eba5db9fb58b674d09b5ba05c8b74459ae8d08234a561ca87665baeb401740fce053bd7574353c8662be0c48e67acff8ca07330825882801a4c7b845213f43da8913df086b48485046ea423fdfa55523a1930f283736a7a5614dabd3abd2901227bd8aa3331eee00a3fa3fda70e91114e89deee858dd5d199c566aa005af39fcb96def00c77ab79b5f3a48e2d679c8c2f00e115211188726e1d03d5173407f6166b468a4115599b613604008e46a909faf4b6587b55727c103f166a907928955988bc306ca654b250efb6351e73d5dd683a742a1749fc64080230f9892518471eac0600c1d6bee5dfce0898c503c455c5f460564edef7c2134cc389f19e1c4b05890458bbd9c9a769ee702e796e968bc3b6791b4c710fe9cefe24d596f3e4888584219bc28bbbeeaa97cc58b6c38f01531ded1fb781a73b82bbc4ac0fca9e005205bbee81197ea88171eacf6b65cefe01effeccf346c887dd900edbb39fbef8ca53e9cd734a7de98efe9893c72589971c7baa00ce100070fc6a905be104f8a168f8238ad6cd21eaa5f5906aa7000547b2aaa8c2f55f1c8dc81b25a87f16acd3d2711e392ae25c03dbab151c9a6b99e862c0f8da5851d903a3805c03368d10c74e0e650596c26c8829bf5b37418f2b7f7da24d85f19594d808a1eec88827653daf5a56a263c42cd9322d4f3fa37860d80f5a3ba0c4b9007f2355c181a383aba2b57b892fdb6f366f52f8066a5a092a9b123472044661be47970ddf3ad84e1b72b1f47b97d42420fec3f50cdc02f9199b83359510fc684f7e70485c6ddc9624b86a416411bde08f402ac8db8ad7ba15558921d62a59c968835c7ac40d72e74ecda85fa9ea15b9bb992339306b6727e559e19ad62e62d514f55fea9d5ca53b2547ac4842f14421e876aced3f1ec39e047541ec38b5e3595d4c272179e27eab4dadd8e0ab2f7789e410ec7c3f2392c478d6fd0373a1ac956ca218a6684fc9c37cbc644b7b3c0c3731e343477e0301d7ae5f512cfcea836c93f65d80595b92452ef7065c924a0dad1f5adbad200c2014b3843e807f77b3fc369af645f25673ab6e874345d47cc61e5461b537565fc84c9da6976bd404755deb569e696296d372ae7413de74deaaab241f375d32fdaebe1b73bd361e92e1d09838f0f4fe96011b403837994666e7ab5", @typed={0x8, 0x7b, 0x0, 0x0, @pid=0xffffffffffffffff}, @generic="cee42dd28e3f52aee8d7a91f41360c271dc994d6c09730284216ae66b25938a87480ed76f5bb06ade0"]}, @typed={0x15, 0x9f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}, 0x1424}], 0x5, &(0x7f0000003140)=[@rights={{0x30, 0x1, 0x1, [0xffffffffffffffff, r1, r2, r3, r4, r5, r6, r7]}}, @rights={{0x2c, 0x1, 0x1, [r8, r9, r10, 0xffffffffffffffff, r11, r12, r13]}}, @rights={{0x1c, 0x1, 0x1, [r14, r15, r16]}}, @rights={{0x18, 0x1, 0x1, [r17, r18]}}], 0x98, 0x400c880}, 0x80) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) (async) r19 = socket$nl_generic(0x10, 0x3, 0x10) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r19, 0x10e, 0x8, &(0x7f0000000040)=0x4, 0x4) (async) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) [ 532.486024][T30288] bridge_slave_0: entered promiscuous mode [ 532.491260][T30288] bridge0: port 2(bridge_slave_1) entered blocking state [ 532.493110][T30288] bridge0: port 2(bridge_slave_1) entered disabled state [ 532.494896][T30288] bridge_slave_1: entered allmulticast mode 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) [ 532.504082][T30288] bridge_slave_1: entered promiscuous mode [ 532.546010][T30402] workqueue: Failed to create a rescuer kthread for wq "nfc3_nci_cmd_wq": -EINTR [ 532.548084][ T137] netdevsim netdevsim0 netdevsim2 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:08:52 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.552332][T30401] workqueue: Failed to create a rescuer kthread for wq "nfc3_nci_rx_wq": -EINTR [ 532.583358][T30288] bond0: (slave bond_slave_0): Enslaving as an active interface with an up link 00:08:52 executing program 4: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xcc, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xbb, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec624", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xcc}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.608628][T30288] bond0: (slave bond_slave_1): Enslaving as an active interface with an up link [ 532.646214][T30288] team0: Port device team_slave_0 added [ 532.667286][ T137] netdevsim netdevsim0 netdevsim1 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 [ 532.682503][T30288] team0: Port device team_slave_1 added 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1}, 0x4000000) 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:52 executing program 2: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.784634][ T137] netdevsim netdevsim0 netdevsim0 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x90, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x80, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e81"]}]}, 0x90}], 0x1}, 0x0) [ 532.809527][T30288] batman_adv: batadv0: Adding interface: batadv_slave_0 [ 532.824085][T30288] batman_adv: batadv0: The MTU of interface batadv_slave_0 is too small (1500) to handle the transport of batman-adv packets. Packets going over this interface will be fragmented on layer2 which could impact the performance. Setting the MTU to 1560 would solve the problem. 00:08:52 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.835459][T30288] batman_adv: batadv0: Not using interface batadv_slave_0 (retrying later): interface not active 00:08:52 executing program 2: r0 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) r2 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) r3 = openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) r5 = syz_open_dev$ptys(0xc, 0x3, 0x0) r6 = openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2", 0xffffffffffffffff, 0xffffffffffffffff}) r8 = openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r9 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r9, 0x0, 0x0, 0x0, 0x0, 0x0) r10 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r10, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) r11 = bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) ioctl$SW_SYNC_IOC_CREATE_FENCE(0xffffffffffffffff, 0xc0285700, &(0x7f0000002f80)={0x10000, "deb61e2f2a84540b872481f48e71267ebf181545a2e58b1a73b1d5c64fa91704", 0xffffffffffffffff}) r13 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000003040)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x101, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x0, 0x3}, 0x48) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000002fc0)={{0x1, 0x1, 0x18, 0xffffffffffffffff, {0xee01, 0xffffffffffffffff}}, './file0\x00'}) r15 = openat$autofs(0xffffffffffffff9c, &(0x7f00000030c0), 0x2c040, 0x0) r16 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r16, &(0x7f00000046c0)={0x0, 0x0, &(0x7f0000004600)=[{0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}], 0x9}, 0x0) r17 = socket$nl_generic(0x10, 0x3, 0x10) r18 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000003100)='./binderfs/binder-control\x00', 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f0000003200)={&(0x7f0000000100)=@proc={0x10, 0x0, 0x25dfdbfe, 0x20000000}, 0xc, &(0x7f0000001800)=[{&(0x7f0000003240)=ANY=[@ANYBLOB="480100003c00020029bd7000fedbdf25c80059800800b60005000000fcef61d3255f79e749b74989cefd2d67f4a3a7f5d872130e436ace7f0251338b726457f2101bcb86d236c5477fb7149c43d955ff467b1d1d809fe6128edcf1b58e28e82300c5c2573e2dc0ad416e2f2871b5d459581f24b06e903ecd8649a80346f46972ab1e0e864198d13853cee89e3a595105e83061e0793077e3f466228c08009600710000000800af00", @ANYRES32=0x0, @ANYBLOB="0d00660080000000000000000500100000000000150098002f6465762f7669727475616c5f6e6369000000000c00c800b8000000000000000c0005000100000000000000c866ef508d669a67a13484453afb67d337f7230ab1bd797f79bcaa1764573b3e11364e29f3db8e20a98ed041912cdf40a5c23af50e2e3a7d2e9315821ce9e21b034756e42c2eb5d5f961a7cae21fbc08bed3e4ad7e2847c6834e35f695631a9b6f03070000"], 0x148}, {&(0x7f0000000140)=ANY=[@ANYBLOB="18ff20471b34cc000000200200022dc77000fedbdf250800e4065f3a788261d9ff0f971ca11525fbe80a9819385c78253494efe3e46d60cc422d35bb9907341cc1a20139fdeb7fbcb083617b8a49cc448c78ba24e08b35f47df84e534deb9226311e8b2aaef55c3a6d1987a75e03ffd476638d92a519543bcff12c98ad38da7835761dd0cdeba44f4d8e9c00000000e64daefac1c90189ce5bd8cca7dfc5689f5c6e6e3cc610249e972f824fade4f3920c46ebb50e3fc50791edff0fcc52b365e1da600d02eb7f000000"], 0x18}, {&(0x7f0000000400)={0x20c, 0x10, 0x100, 0x70bd25, 0x25dfdbfe, "", [@nested={0x40, 0x131, 0x0, 0x1, [@typed={0x8, 0xf2, 0x0, 0x0, @pid=0xffffffffffffffff}, @typed={0x8, 0x9a, 0x0, 0x0, @pid}, @typed={0x15, 0x27, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @typed={0x14, 0x58, 0x0, 0x0, @ipv6=@local}]}, @typed={0x4, 0x64}, @generic="c789e7cc7ca8dab63be9ed69e6e9a202e2d52f41aa", @generic="bfa355ce6fab563c093e78c05135747bfff952128f788a9c1ac249a7191e52772e6dabd2567244773291d880df05c45846fb4f3f987662064e2bb43b7ed60d3c4b6e54c4409c3ed0d40e512cbb060d95117797bd7f17b73f40f7dc49ea6c0be1efabf891a76d6f8550dd5d4686af8f9075a2c165e3a108c2c6f279672edfa2e65ce7a96a1ae680700aa7fe2a58ce", @nested={0x114, 0x105, 0x0, 0x1, [@generic="924193881425028ded5e45af38309971b787cba9b5eea3c02361f21ad6ec6bec8f6a34d2be156296c42513da5533fb9d3af4038f7b02c9fd412b3b70f94bb65e569922e3e9f346c0b8ed2ad109e0d5f3b953e5411ddfa7ed7b0a60542e24b0510f3d1a96629ba0a58021eb623f3acd60ad8f79c4278dd5c9213020d6359fb392fec783bed11fc3484972230b510f0e63cadf00b0b65fa135c61fe0cbf3280ed4a745719d8945f1fa1864a507ff66b293b5c51bdcb35a527746d528fb38c049829187d0a677b124fa68b2d6ceb565e1321b86fc", @generic="4d721c8b383769a764937f17db741ac512ad5f84dc682af8edeff4a74f6090876cc97ba3b128eecb2eef11038b50a328d818c4776ad01890bf08440a9a"]}]}, 0x20c}, {&(0x7f0000000740)={0x1084, 0x10, 0x800, 0x70bd25, 0x25dfdbfd, "", [@typed={0x14, 0x5d, 0x0, 0x0, @ipv6=@loopback}, @typed={0x24, 0x18, 0x0, 0x0, @binary="17089d9abf899777e6a2d77f1e7d5724872ee91444fa4da34531d08dc7f9ae85"}, @typed={0xc, 0x128, 0x0, 0x0, @u64=0x9}, @typed={0x8, 0xe9, 0x0, 0x0, @pid}, @typed={0x1004, 0x25, 0x0, 0x0, @binary="4d25059a1a01986681c822bc09f098bb73c119783f81dbe8824a8fa9014888c3f670af704119f6cc1545d35174c987fb52d9051bf3b01e4c7a424f595280a47cc3a01b45f06cba1defe1ab7ce1d9f66c00f7e039720b0dfc9e1d53cb304618a32f26256dccd3da1d40e44d66bf9f3c413ebe925e39394fb082f7818758d0b7e58df5069d078aaf74adea049747759263eea91b63f86a6ddf0b3290685b897c6073700c5c1dd8fea2dd1d01602e4299d5f125cf0163fa2c1ebc6de3f042eb9bde2e479ebd7a738f35c4cdb587bef8e571e0edaf3c4705553180e1cf9bf0adf27fd2835864e25fad08d8cfaac58e6c60d720aa7903db9ccb41fcf8793e14f97c81867739fb11c90d583f0a35654979a5a00e8e43e542077da3c7d6f7466d5a00d5d60a23b518e501fcd3e489c97762f6173cbb973c450b3b26edeb7a730407c74e7f11aa25de3bfa907086945e67b54a837c73567ae62633dafebad4d38ab5813bdd9d80262b6d2471a1a61cb97dc4fcdebf0cd980c04319b0463b1d0db04131c411e535438233a2028650ace7688bec3de1806e44ea79a63aade453acfcdbf5aa3e63c3e075a5caf2bc745d88554b7d23afd40094c9e654a253050ccc8e1d549c6042c9f8da7cbea3898c8573b04b30a678bf6d39c0f68fd2cded90da3dc4a008dcb455a30765aba41fdf7b9324d7ba34ac8dbbd47bdbd1b022b78ce2b765342dd776059d3a79d86b317e199147b1933b0b1def209de8492bd6962fbc93338eea9a5b3288161294459c765bb8a982f8b16f258aec9efbaaf651667758f05a7e11163dcd36c75bccc3a6df491499ec6b5c9e86dbb37a495033bcf5fa1a4ab5de98a8810e717c7f8db4a22076ddc32271f4fcf615c0a80655a7bb304f1d020587f34e6e7c1b99b6102ef51a8d92d7373d849e170129f4a1801d45d5aaff670c07afddfdc55fb0a597951c58603585f247f5a11f9c35ec83a1680f103f8ecf6deead4ad441cd67a4049486a16ccb6ccb5e935bf3a002312a2c419a0b14dd436730bd9b7f5c74c36f6b3dbcd184d750d0278fb1f8184d0e7545b7d7af48d71ca0372ea5b8e9d8ad797dafce2418bb24513a2653a7786998669795d8146f5b7ce28fc183aada6a942d1e0df98c1dedb579a46ed84a97798e6c4415266c114c659b01cb35c88a6c4b4b0383c5fc2b06c551319ebc9040a9d2939cc8bc1f1e5d0e156ef1d4d782ed7d87a28f3ef15a4494ef55ec26eb665078fa23548b9a66e438c3f747ff20ab2ea6aa03298b45fac67ff91dc3aee6abc127d5ddfd1a6f8d91999f85bc95483ddf64ee31f0a3d93312b5b29083233478ac85e37f9550f12cebeb072d24932145c7a211f4420f2bcabc52cb4caeb92354021181554ad0a4ae1c0d9d90707990362236d42cb29147f27d6830a4123d0ac346ee69a91922107d80b3457a66f274d98ef2d867022b34379d229ac4e702a64db3bf011bd1b4daf0a6d878ee9b681c9f52e5560a20225e871fa7ac880881f1e39a602400ecfa89665bb2558181232ded9fdf3a1063f4a1ff36241fd0ad58f09c05471dfc263a52dee36cf97f0915c22e7b22abfece46d5b0650fa1517a0fc3f820d5c7b378256daeee7723d9f04858747d630d3fbfcae75210663075676d890c492a50a0548f039deb20fe0a33017c1c721df39883dd6484267e05fb244370fcb729b522eee69ef47ec7934a380531a5f42b139822b6c7cff00d5cfb291ba9e2edcdf8b8b0735e49ccc2dabe9a1c363feeb821b04a645d3686544a212e887aba2bc4b6f171873c465a8a70e0810f036a73eb86ffc67182460373fb3a1b858eaac511934177126ac3060b11480fbc4474b558c496a27221a55d72370b498e081ed17d1197a42b373ccb86d76a2487dd66a60acc61bcbe7f61b72bd9e533793b7afc1425f1e8133c791cd65b4e766642c9497fc5ccca4a530d0235a620d2b08e2bc6e2b940d7e2b6312d489f63d525ea2e8c4df4d4ac05435b63681b91b13ae0d203eaa722662cdc33b8ac68fbaa2ff7f657952dd46650bb09633c519cc7c44126ee05babc55a3fbc2933bedf7cffcc55d7a3fcc4460e040e2896d1d8070246c821eef4e6723bc1cab07b305a4fb4ac6d78898d12144f56e25c8c5df9e6c80c7f866c3961ecbbaf1eed026da06b776c5591e1af5a7954bbb4efb0c1af303b688aff0576b1c50b79e6e310ba020ce3dcb564dd606d650a3dc2716d157bf013211468536ad0e6646213eb3fb02a884dac52b54a5aac329ec1eb4a5fb2beaa82f4a245410b3e583cd3c0014fca640d7a752ca6dd47c271d1d4ada9bb79b42501cb4f5894b9b9681a43bfae67b234288ee50e0101decda623569d5d9dad0a7e4c608858fb04de1cf356c9f0db9ea9cb5df287fb23efa0b17082c1257d77c16539bccfd3e7591540eb82e9f765dbb5bb289f2bd489a0488032b5461d1348d670693f4e53341f2456ace8e2f59ac470dc8c999e9e52c2e0e320a0e2d1238296d1185561002d93825b147935945c6d0139fe99d3a155a5c1afc2f3f176dd521fe4927737d83d33ce48d333f39158f5653f8345b7cda03ef8dd217868e7b8e06ca3b771fb923458fb649aeeca8e9f08ff2db3872ffffa6e7f773c37da792fd1c3030fe94f1ef0b971b4b6632783f15dcadfaab99be10db107957061340730f37c1c5655117dc0115f33987e051e6acffaaff9491694e4de23d08a24a986c1e866dd1910d215e22b591069971355cb145f52fc22caf8f5b120a9f4f9723cae79e1d15f05a2494d730e31e548148e9b326f37d9565157d03415b9f62c631d41c7d8d0b40b5b99f02dadc6ab95d81a97098628cf1c2803228169e4aad29f298753ef948a89c9dd38d4e953b0f3ec6d97a33d49c097baea530ad9f766d51f695976398d8e10f8c3fcf16b2bdf8f1d2535accf1fa0e85edbc4f189dc4ce16c9a7330fc77864b67a6b296b1349efc28d4281320214024fa851f467d3acb8d3b56ec98ffc73a330d8c395c5b2b264b24dcd6d03d4a7f558b30be5b22c6707b42e855a08df63825a89157bf76031f574454b142f05fe90bc9eeee2d7669615aff7233e6c01dd10da16a0fb5b488429e1fd936fc60b4897bc62c708e6df7d244e1631879bed247ae681bf434196d8db50ab1f702775c567bbc7f811fe4d00682f1d0084f67610318bee61189ac300da5cacacf85f067b5b21c6d68cdd3d11d48f365c7a187c2895765131800ae6d4b64c37ab8211f3d870284edbce98669e6721969e5690876f6807b9233784bc58a50cc75350f21622cb9e582644571add1159c307e0c2d385eef4b92faf94826b52fe7075ee40d891114e9ca69a389e1d571cd4c66bee43876b1251828a1a181b4011197c3fdd2ee610dd463320afc1e2f17c1f23ef6b4dfb1a73dc8747c94060a741cabedbad0114bfff358d0508c977b784f0d352de6b7c1e49dfe47e99211483be99e7b33147394f0fd9bc87f3cf0c02d174fbb09ff70add88ac6d5e243245b09d8e2b0d349698bd38f75ccfbd989ceb1bc4bb5732eddb177f97c80e0438d5e1ba15c9ff2ff50ba19ad93b4356a2664a73c8cb33663097887d7fe88c1de3d78309ebe7a16f483f544406ac448aa9356bd9171e4aa027e43295bc260bb1b6ba1fb00a4ac4490d1dd47e84bd6593231e1ed6e1e4d1defe56d7843e5a0120c2d33a9303349f424fc4476f27e2454edf30954496e5b4e789802eae2229f43073ad5c1ef64ffa3f8b4d0219b8812f2991866c36a7b1dfd8b510f69d1d1549d26db9d3c3cf7aefe1665c81082f2b9fe58d0e7670e9fdcda32566cb396a5ff81852b93945c6720554f36dcbc79bd44b5145d9e8382d6277afa4eb6a28d36f8239a533be4f9ee1fe8447ed63909391208b8b3af9df22e834ec9718f3474e73532fee8e113066d8eb5c317e97b56e7aa0749b4d84d8dd5435000c05e6b73ad4978cdbcc47480737923eafc754ea4587ce6b9b08269a3a73838e93c226d50449eeda62858617f98846c0658e9ffa8ecd2559b21fea3db08f413b593fa8c3e09d572993fe7acb83da43d75a5a789e5bca383f32966271ef7d4c1c6be53b98a1a7ec2c23957cccf019f7a7c0f9a87d89f1c15bcedbcef2bc4ce081aff354bc3df005ec813693348af7a4f6fa81ae4ec3616d63b07e1397cab03dc36654503b38dfc02250e00ce7e6d73ea53437b467df05979636a9d50e0b326049db5e801c4925ef646d85eb02b46d40a596b5c9b05cf31f441b97afde728d4c4c8062603f18db5b60eb327bc39dfe963601561728d4537c1da7a6120204b8b756d950c8779aec13c1891e0c1dc090857d3a5c15c074c2c86698dd9dbb43b8d789498c500058a3ac00b04ba5b4c648a7192751fb1d32d40786774ea04ad350ba349b7c0e37c39cd25d2e93d9ff06b0be3646fad982ff23f4f98fdd724396b641746ee2d8f786e7e81b16b59f4e5990ec945bc0294145e63295618e6f94ea3ab91e8bf28e65a395a27716ee8cd93d8095ee9f8472d879a383290ba00d74185bd6a34cd205eda9c4812f37eb1216cb6af28c87aec4771d7e2d9f52e525a3cc9abfc72663e5d791d4350b4393c7cb5e0572eb96de78389d794f80ebd9bc92b49445f6b68c52e6c5a6c68cfdef5b7f59626fb2a8026b906bb8c0e86f62a364c16ba1bb9ccc8f8288cc6cced7e1ecd974090dabc3a517634ce63d589b7256398037402969a485e92ff6fc6d9a572d3b9d423c1f76ac1de491792f373aaa2bc89dfb343f321b9949918504f515897f0723beadecb233327cb163116eabb36bb6543cb0cb3c2cb1145797afcfc6d626797b28f5f03bc2cd1d7f555283233d50dac1dc1f805782952e8a34c0951ae8a6691c31fa1eac2bed1c33c8ee5596002b3c687ebcee2d47475276d1cc1627866b904f5c3330fdc47061b7e104b4c59a5d019317865e71c347661f0aaa5035cfb216bcc3427fc0cfe06fc8522f04e7118584183e8722dfea9bb629a365d0b77e0e4689598a407765da16483309770b1186262f9625d507f93da36f2cda11c97c498a3cff11d6f351346f882063a277acf0f80bd452e90a44c161893b4aaade86750aa6c1ddd6df3254e1dd4e3a18054a369fc9e98ebd81d7ec35bfecda1a6ab66ec8c63331369649446d9d0da3ab9c288b83c0db15863c7aabcc68c4249b8cb7d12b46d9baeb2752ed26d71ffe48ad6392693153b13e7c8cc326c12b5603cfc9f2172269f547d2529fc70291bb38e3bb3279b9d765969bbb62b0b905251de86f0032ec26b502e907ca47756acb5bf49ac9014690d0a112d5bcf748a0fc1a6a65e9b69280ace21e9b048d4c519f0bdfa4639c5075e1cd724f3483056bca45f273ebc304ad2c26a651c31559accc7170c4f227a260fcd6127d5283dbca6241631d2752902223ad84cc715b219ccb2cf7ee1f1624fcc63739d828c543bd2ba6353f784efe7dc0e9c7837d5fdad8bc35849b18299eecb7936231181a87a3db380760c98f02389c37c0ed6758ee80d5497d0e1b2f3367efcd0225cccc3a09db7a8c8801ff8dcd86d92b53def8b3a61dc00de3f151628603238fad3b1c1cd38d98a0da9e30ab6e7228eb9970549ff06391adfa918f1257a827626b9c304d0687d0c163c3563867c317cacd371313ca74d0d05326546e9df2f22f9b1c21de253991321a13452a3b800c64e0f33c4d9b232058ba40ab92b2010464b4ca8fc10de7baf4860d80ce4b948ef95f6c684cf3c061461eaf5341c09c596eeea339e626c2f4f7db554283d590fa9b83"}, @nested={0x24, 0xdd, 0x0, 0x1, [@typed={0x8, 0xa2, 0x0, 0x0, @fd=r0}, @typed={0x15, 0x82, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}]}, 0x1084}, {&(0x7f00000019c0)={0x1424, 0x17, 0x2, 0x70bd29, 0x25dfdbfc, "", [@generic="133f5ce256e48e186b57eb92e3d67968193a2cf400d7cf3041790bce471bc257c3438d3e71a7d8393ef1b223f84de683314deeff92769a9fd838660088a17f6c2d29881a441cc1f341dc6ee6f4c5844598db7afd982706b468a39ff3ea4ad1631488fad356047ccc217c45ca7627ffc75733e22f495b0f256e7210a8d146509c7af6c38561393519a3b182b78ec1dd9bd25477a3cf587b27", @generic="e88ec576b23c16f99f3075597fe305d7cca55f465552bc35edd73b9e4d5ce2b8b486a37fc83eb8ac46f197789352d0b11938a7ba84027eff3fad4ac1222bb0d37e0471e0171b48d59270a259b2744ff58a8de6debd54f67c78bb47fbd8121844e292a0cb7e8a95c97e149a4d8f3ff033d6975eac8d84cbd3bf146fff16bc4690f3f3ac657ca1b672ece1672fbdb017c389dc634bf6bfb31190643d731343196c3facfa4d7450b8271a53706e4d46f5f873e479f3e1f3304d18c78039dec6ce750ec093f530b30b046597087ca6a4cb344ced7bdae4bbbcb05d3ef335cad3a84c072e397d382e1a4badfd816f22f0e243e0fcf79bbf3a6ce8f2", @generic="143fdd42b4d1a20607d75360dc83c395c8c3ce8252a2ce7008653d1b5e53fb041a8ccac0cb52b2df15a1e74b2e4c43a4ea411c43ffd236f7c089da121b3494d15fb915c4a8f58de93e5a92789bc38b542602a9d8af3e242d4a2e", @typed={0x14, 0x7, 0x0, 0x0, @ipv6=@ipv4={'\x00', '\xff\xff', @multicast1}}, @nested={0x120, 0xc6, 0x0, 0x1, [@typed={0x8, 0x3a, 0x0, 0x0, @ipv4=@initdev={0xac, 0x1e, 0x1, 0x0}}, @typed={0x8, 0x75, 0x0, 0x0, @u32=0x401}, @generic="b6895074bb7a3afe9630c6f99260b1f06c93bc2c0722540649440807d7a24b97f13ff6d21106215013a3fd5699a435b16158e53afb0f6ae0ff043d6639be1fb09a7c68", @generic="2749fb2104172d73bb85cc455515bcaf861c1add239821475b56e3cb827b64c9e8c2862d0fe2709bf0adc4377a55e79ee35151f9038a77fd6c62c16c9c743284915f107162cfc72596d29436f2527a36488416958a5c68125d51c5144522d271b74c6073546aae75103ccc5369d329bc92d13548f43eb3ff88fb916774b5d66961283b0fbbb72316c3552ef5507be8157459bd55be305195f50488b1a1", @typed={0x8, 0x105, 0x0, 0x0, @ipv4=@multicast2}, @generic="fd03adb6efa816bd7ffbfd7b56324c6291c0d47943cb464325b3037bcd446f2e310f70b6"]}, @nested={0x10da, 0xd, 0x0, 0x1, [@generic="845dde559616927f6d889da26095d1f40f2712fc6c56ebc6a455054ab7", @typed={0x8, 0x82, 0x0, 0x0, @u32=0x3f}, @typed={0x4, 0x50}, @typed={0x8, 0x57, 0x0, 0x0, @u32}, @typed={0x4, 0xa5}, @typed={0x15, 0x10b, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ed4141bc4f65989a9acee3dac370708ba95194311416f97e58f4a032c20ce1e20cdda03f81f0af7eed5353f09df66fda0010f3adb90b890776cd0a4e3436e5be3ca67c868e0ac0d2afe1832180a82f908afcbf870998621b", @generic="035ae798dea74a23cca6ba07b89ba8e528f72a786258952651b2346be481a4d62a955679adbf5cb2c0a16f7a01a314546c02d732e756dfee390855b0d51d78a4f4725565deafbcecbb65ab1fa666fef767d6019ed11646ccc00c84e249142e19a026cfaad65b4d2180b754ab21aa2d8c6c89eb9b2de191e769a05af2d8dc62f9bc80a546c4dd13b1f23f75253431fe2e92c09a577c7ea70dd8d9a348ecf6e7ccbea1c0c67ff07ba9edf44e3b3e50cc631d175fbb1480363d63c066920b56209b292fd82d1f11c66b7155ac3740daafd76cb43e70aacbcf63ce4e63604b6f9b178a6bf3dd0ebe5261ecb144838673d88e77eaa8686a4f9bb7eee95ed1e5698275ba14d32366d90e12551165017478e4774d497dba0995f4aa962010697dc9d7743991ca5334f52acb6f46f1e76a6c7d353d9f4b2de20ee797c17c599391c5f94249bd7dc0475caf9c43f839a745cb210055959073d5ef602d12f1bf8db2c7e60093f5fbffa27ecd3575d6ca03dd468f854cec88606babd590edc6fa795f2525fad1dcb81cab9e980739e1622c9650d335f8f00751d87bfb980fd1940c3bc7b06bb73062faed4e73b635ebd5109ed8bc8c21c4ec776345cd488a59617224143305d9fdcb064e65b9109dc2389103987ee41468d1613987bffdd8bfe7929e7d2d799fb22a9bc8a400abb17c962497b5d6579c55db7979b8a60949ee31759f5013873d5f610056ab9057065381f86cb44c5c392a855eccb76d40276b0be45a6919163a5a2a60c7e735332c87cf8f56d01dc8122f01a55444e9d8857acff85d3776cb6973840969898b6e26cb3c2b4272481ac3f2ec0d4a431c37a6db5b543af703245b3d2e6f1a1f122cd4cdf978285aad0ef7be6e5e5738a41fe8a941ae84811dba1ba265351fc85d95325c40fc1d737992fca90fa0738b8cd2108f9238490ff4c7a84d56e6009568b6febaf03b723abf1742b8cf3311331b33c3b9cd1f25ccf25db2f06ec2d2b0664dffdf05831c51a1a08fc499f9e8e04c23b0f7b5106e65913a65baaf963e4d38b13c4de41cc39cacc0393909a61a6fb1d98455c4831b7f8295f95d1b5b6af46193fd97296fc66bb3319f9d1b75818ce9f3176de68d4f8c4f758b1157a4709222df95d3a34cbe094086cb78cad42b974b03828fe81afff1e8a733060d0f63a6a5d589cafd566b36c6eeb512dbee5805407a57859d2195c5e8032f961552e7585671490cddd0cc1e6f516dd9f838040fcaa94ce4367642ffcfe7a0c112a25381031465d990ab1a01b75a5966d04844065f1f118cf0d1c116b6c34f796957fc2228cfca4936d8036b0f4a5d39c12bd660c37aa8bd47a303583a520d3db871b2de6990098dbcaa66b5513e493b50da6ea8415414f95965a8424c727d48e1449780e8860513244b61429b8a2a6d0b881239e54882d05920e4730863f19b34f6f48d26e7c45a65625cd80ba359e5d47ea0373deced5f0a4e84f9718d2e6aa46adb3a9c87b650ffe925cd7063de5d72912c308ffa9f7d6641c9ec82e038f51493792b42fc8019b480d0d54985fc0f7f86c80aafefc4bb3f9da67fbe034d1713319ae95c47ced97e09d5e121212eb4f5e10a55987edc54f13739e534a9ab9839bf1f3a42f6311c748d465fc8e117a26766e73d59898e34d2eee291cd20cddd6f3ad1d07586eb4db733440a3179b53c87b39a185acfab259e70cf078f6b878ba8db4ae98d9748a13670feb16731117b9b10a957fd51394a3b412c836f6ce1af7070a2cb3d73b155f576d821d4491c8de854d02353f815974b6d5a560944d781530c1e7af0e302730ccb5aa38b4b1cc1cf452a0fc0e9a091798b2f4e35be94c8a8a72f07853ccbf3e226a0259a0051c7638c4a285e4d7f6b606e4eb707d51d8ef3b1df65e1cb48849a29de75adbb19429c2e022a05c751459cd0ff536b42e2186aea7a299ead87ee2a050859b05da1698c94be83dd74b4b466cd962557ad32865966732c352e527fa0b0407a6b0022766c64566a3ec4861937747f2e07fce824b2ca0d8d12e3538c8a3d25a87f8332a9cbc83f124bcc0b886f4cf82c0d6110cb8f6f0cf18aa3dc88f66a8e3f59fd4480c3adeb1492da7c7f177870cd9808f3231eef91ad6b622bcaf55643e3ddc8d3b1f8b57602eb4fffa127b78d8fe424c385e4e577f8f31d424c6f4fd86aa0667d98ac2a6de2f598a14a160f2495f69888d432812493a0ca775d3cdf40818c7ff9e5a2164f715c993625588a68ed971849330cfb0d32ae6360a3b7c2b0ea413242291706ebc4ad649ca1a57e9de7d99f574d3891bb2912c921b7a7f9a90bf7c428b6970b21b8158b0c913caeff7ae6dfc745d367cf5b0b93c4b39e8c9a0788360aa41785960326d4abb8c16a9dc25aa4f63609e5909966bede3a99084ac9b3040b68aa7ad5bd7cd79e932931bde3accd9a46db66b19c5919a68992452e1a4d3b1506ad99c4a56f66f8cdc02f50c3e904957e3fe3e72d0ac3f6f10c5dd838afafaa5a8a052ceb37fa97fdb9d8e71004aba543ecca85ec88d69db7d46bc1bf089a614a397ffed1c89f43e0bd3f5668d628f22574f713fa793d3ad8ba14c08bc7071424e4dcd821e5d914b29e6862afee6e2f695544e28de50dee2c9b31ac3d36b4a53188dd1f602ac3faac101218b92e19b396353680dba8b0d7a86d00222832f560269362a07c12b25b9a42d748dfc98b8cd05ec713a7f4e17fefb7ddd49a7a38a72bd11d8f2da98aff2521ed44a2dda6c5fe56f7eb78ce6f5945e2a0e04b7e74e692640c109560b8037a0b995bb50b951c765225a3736a03c5c999328a6e404d3627da9aea8e110ece0d57296a86c780c16e6d21c6f9d3c1e6997e00a9bcc40b7b02e3964e25f291dd2e658c8cdf3497d35c2597923fcf8a87223bde8166933754710b2cdcd0b9935f0c04cc97ce201966f0b945708875a428394b4c7faf661ba8b6d35f0ba5d4f6bd51502ae25893ba2b32147689448410a40a09e14b4077c11f271fcb46cddce4ae16e329c09061f9398916ffd6137e6d106afbe97db920abb260c1a218767449fbedb7919518c1bbca93a8b7f640468a0f987a8258d55ec2702ee7ad386f412b07d77920ca6cb2c36cb426de4f3e90092db6e8d2d786d520d0dc0441bc98dc6a10fec47f87b9ead5c11f232035225f826fab74b4a45279344cb38174ddb9fe7794017898ac714e0892098fa9f5f522d6c16c9af130f3bcd225d9f40324edcd7e7aa8031bc30460ba4285e8732fb5e3d0cb7736b05540e340ff7ed7fa540e166d1eb053f9c22e342e7d30d7bdc836ea5f1df25233e6e50a3520138a7eebe0ef6d72e8699b0f68b4fc99e899ba8d1669cbe7c9af296391083e15b5374162c03e1b0e5de3f0f250eb1d040cf2a4b8cfa4799c8281a4e8c92bc244730814d19eb2c14ecdaa525253e112d12757d49d92b4ae22b502781d5c20d5dfaf6f80d8db1a5b3dae00dd3146438e2cc573c93a1d7c3057a308255d002057f5f491463f7a018bb63ec9008b05a089b70512851c6449b4127efee6a8baaf4afc5581c49c48d5bb1db182bc936ddd0505ab7a941bd29b7e674cdf4c3327aaf53364271fd79f7f21fce359de940d9a82d17bd0df7cf25c27c45c49e3664178464ca791f809906de74bd710926561d5ffb651a37b6dbf12db36b7e3de67027b6a5557cac24cda44b86d741bbdb9ae65cca48d355077bea2d606f0e91357d47430fe7c7ee8df80a746fa4d553b089729dc9b3a4cc980314bc53e4a3c65578687408f1038ce7571630486a14f187308e291b3feb9799d4642dafb99192902b86e8ac8a487637a7288710073e0e4d0b1174b08f326f9c55debbf66468f4c1da90762b64b91b99717022a463bccc9b6bff95ae10cf46759bde811f6b22846b9acb152e978a82afb91ae44a2ca1094fe6a757adea40c7705e3d2e5549733da91f22f4e34ec59b3045cfda255e06da9b8e5eb17a95a43775d67a9add8e4b530656a2d6a5ed5bc8d78595d6d2688dcd09d51173a6670c7c71b6913dc92c08bb4e4de9522b338c581356c37a7ddaf509050b50b0e74e401c0b76fb0062407664cb076808aa59991593d453c1c7e090bac17b88eb6b5e6f0ae9dc35173be30fdaa7cecd8285a7235a1d8979e6a769143fe938bb4047167cd01055fd080f7d8d435fe7d3e03a53e7ca4598cb3b4c265352891b6b40872625f157ae43d703aaf9c4376d9f02c003f3488b90378fa1c4936829c09c70c6fa8f049883c909fadc7e0b492e2823ecde921c6050b7f36dea0047a9a9bbf837a414d46bcaf3fd0faa4422ef11f4054294d25e553203f0f9354756521d4805c14a7b694d0808c55455e06cba9d93ce10b3ba082f25ff35feca2e4413605f3e470715ad9f1c10d6feb246e848b4d9c138fe734640ff125308e157c1f22c2e49acd2ecce153d0ea9ef4bcfb83b8faab3e729c03d85f3791ace97ee6fb594bf8510576790d4c996c0179c7ed72ee90ffdb5e47dffa3e0baa9d4bcec2630c51ec8e7afeeda361eb73d5215a70ab010ef60a5b3b291adb4a183df513958ee46349e290da8ef362fc62baa4ac788dfd06e7b385f04a3f95222a38be90eba5db9fb58b674d09b5ba05c8b74459ae8d08234a561ca87665baeb401740fce053bd7574353c8662be0c48e67acff8ca07330825882801a4c7b845213f43da8913df086b48485046ea423fdfa55523a1930f283736a7a5614dabd3abd2901227bd8aa3331eee00a3fa3fda70e91114e89deee858dd5d199c566aa005af39fcb96def00c77ab79b5f3a48e2d679c8c2f00e115211188726e1d03d5173407f6166b468a4115599b613604008e46a909faf4b6587b55727c103f166a907928955988bc306ca654b250efb6351e73d5dd683a742a1749fc64080230f9892518471eac0600c1d6bee5dfce0898c503c455c5f460564edef7c2134cc389f19e1c4b05890458bbd9c9a769ee702e796e968bc3b6791b4c710fe9cefe24d596f3e4888584219bc28bbbeeaa97cc58b6c38f01531ded1fb781a73b82bbc4ac0fca9e005205bbee81197ea88171eacf6b65cefe01effeccf346c887dd900edbb39fbef8ca53e9cd734a7de98efe9893c72589971c7baa00ce100070fc6a905be104f8a168f8238ad6cd21eaa5f5906aa7000547b2aaa8c2f55f1c8dc81b25a87f16acd3d2711e392ae25c03dbab151c9a6b99e862c0f8da5851d903a3805c03368d10c74e0e650596c26c8829bf5b37418f2b7f7da24d85f19594d808a1eec88827653daf5a56a263c42cd9322d4f3fa37860d80f5a3ba0c4b9007f2355c181a383aba2b57b892fdb6f366f52f8066a5a092a9b123472044661be47970ddf3ad84e1b72b1f47b97d42420fec3f50cdc02f9199b83359510fc684f7e70485c6ddc9624b86a416411bde08f402ac8db8ad7ba15558921d62a59c968835c7ac40d72e74ecda85fa9ea15b9bb992339306b6727e559e19ad62e62d514f55fea9d5ca53b2547ac4842f14421e876aced3f1ec39e047541ec38b5e3595d4c272179e27eab4dadd8e0ab2f7789e410ec7c3f2392c478d6fd0373a1ac956ca218a6684fc9c37cbc644b7b3c0c3731e343477e0301d7ae5f512cfcea836c93f65d80595b92452ef7065c924a0dad1f5adbad200c2014b3843e807f77b3fc369af645f25673ab6e874345d47cc61e5461b537565fc84c9da6976bd404755deb569e696296d372ae7413de74deaaab241f375d32fdaebe1b73bd361e92e1d09838f0f4fe96011b403837994666e7ab5", @typed={0x8, 0x7b, 0x0, 0x0, @pid=0xffffffffffffffff}, @generic="cee42dd28e3f52aee8d7a91f41360c271dc994d6c09730284216ae66b25938a87480ed76f5bb06ade0"]}, @typed={0x15, 0x9f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}, 0x1424}], 0x5, &(0x7f0000003140)=[@rights={{0x30, 0x1, 0x1, [0xffffffffffffffff, r1, r2, r3, r4, r5, r6, r7]}}, @rights={{0x2c, 0x1, 0x1, [r8, r9, r10, 0xffffffffffffffff, r11, r12, r13]}}, @rights={{0x1c, 0x1, 0x1, [r14, r15, r16]}}, @rights={{0x18, 0x1, 0x1, [r17, r18]}}], 0x98, 0x400c880}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r19 = socket$nl_generic(0x10, 0x3, 0x10) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r19, 0x10e, 0x8, &(0x7f0000000040)=0x4, 0x4) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) [ 532.864423][T30288] batman_adv: batadv0: Adding interface: batadv_slave_1 [ 532.866232][T30288] batman_adv: batadv0: The MTU of interface batadv_slave_1 is too small (1500) to handle the transport of batman-adv packets. Packets going over this interface will be fragmented on layer2 which could impact the performance. Setting the MTU to 1560 would solve the problem. [ 532.873821][T30288] batman_adv: batadv0: Not using interface batadv_slave_1 (retrying later): interface not active 00:08:52 executing program 1: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:52 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.947898][T30288] hsr_slave_0: entered promiscuous mode 00:08:52 executing program 3: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 532.996938][T30288] hsr_slave_1: entered promiscuous mode 00:08:53 executing program 3: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 4: r0 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) r2 = ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) r3 = openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) r5 = syz_open_dev$ptys(0xc, 0x3, 0x0) r6 = openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2", 0xffffffffffffffff, 0xffffffffffffffff}) r8 = openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r9 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r9, 0x0, 0x0, 0x0, 0x0, 0x0) r10 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r10, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) r11 = bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) ioctl$SW_SYNC_IOC_CREATE_FENCE(0xffffffffffffffff, 0xc0285700, &(0x7f0000002f80)={0x10000, "deb61e2f2a84540b872481f48e71267ebf181545a2e58b1a73b1d5c64fa91704", 0xffffffffffffffff}) r13 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000003040)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x101, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x0, 0x3}, 0x48) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000002fc0)={{0x1, 0x1, 0x18, 0xffffffffffffffff, {0xee01, 0xffffffffffffffff}}, './file0\x00'}) r15 = openat$autofs(0xffffffffffffff9c, &(0x7f00000030c0), 0x2c040, 0x0) r16 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r16, &(0x7f00000046c0)={0x0, 0x0, &(0x7f0000004600)=[{0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}, {0x0}], 0x9}, 0x0) r17 = socket$nl_generic(0x10, 0x3, 0x10) r18 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000003100)='./binderfs/binder-control\x00', 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f0000003200)={&(0x7f0000000100)=@proc={0x10, 0x0, 0x25dfdbfe, 0x20000000}, 0xc, &(0x7f0000001800)=[{&(0x7f0000003240)=ANY=[@ANYBLOB="480100003c00020029bd7000fedbdf25c80059800800b60005000000fcef61d3255f79e749b74989cefd2d67f4a3a7f5d872130e436ace7f0251338b726457f2101bcb86d236c5477fb7149c43d955ff467b1d1d809fe6128edcf1b58e28e82300c5c2573e2dc0ad416e2f2871b5d459581f24b06e903ecd8649a80346f46972ab1e0e864198d13853cee89e3a595105e83061e0793077e3f466228c08009600710000000800af00", @ANYRES32=0x0, @ANYBLOB="0d00660080000000000000000500100000000000150098002f6465762f7669727475616c5f6e6369000000000c00c800b8000000000000000c0005000100000000000000c866ef508d669a67a13484453afb67d337f7230ab1bd797f79bcaa1764573b3e11364e29f3db8e20a98ed041912cdf40a5c23af50e2e3a7d2e9315821ce9e21b034756e42c2eb5d5f961a7cae21fbc08bed3e4ad7e2847c6834e35f695631a9b6f03070000"], 0x148}, {&(0x7f0000000140)=ANY=[@ANYBLOB="18ff20471b34cc000000200200022dc77000fedbdf250800e4065f3a788261d9ff0f971ca11525fbe80a9819385c78253494efe3e46d60cc422d35bb9907341cc1a20139fdeb7fbcb083617b8a49cc448c78ba24e08b35f47df84e534deb9226311e8b2aaef55c3a6d1987a75e03ffd476638d92a519543bcff12c98ad38da7835761dd0cdeba44f4d8e9c00000000e64daefac1c90189ce5bd8cca7dfc5689f5c6e6e3cc610249e972f824fade4f3920c46ebb50e3fc50791edff0fcc52b365e1da600d02eb7f000000"], 0x18}, {&(0x7f0000000400)={0x20c, 0x10, 0x100, 0x70bd25, 0x25dfdbfe, "", [@nested={0x40, 0x131, 0x0, 0x1, [@typed={0x8, 0xf2, 0x0, 0x0, @pid=0xffffffffffffffff}, @typed={0x8, 0x9a, 0x0, 0x0, @pid}, @typed={0x15, 0x27, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @typed={0x14, 0x58, 0x0, 0x0, @ipv6=@local}]}, @typed={0x4, 0x64}, @generic="c789e7cc7ca8dab63be9ed69e6e9a202e2d52f41aa", @generic="bfa355ce6fab563c093e78c05135747bfff952128f788a9c1ac249a7191e52772e6dabd2567244773291d880df05c45846fb4f3f987662064e2bb43b7ed60d3c4b6e54c4409c3ed0d40e512cbb060d95117797bd7f17b73f40f7dc49ea6c0be1efabf891a76d6f8550dd5d4686af8f9075a2c165e3a108c2c6f279672edfa2e65ce7a96a1ae680700aa7fe2a58ce", @nested={0x114, 0x105, 0x0, 0x1, [@generic="924193881425028ded5e45af38309971b787cba9b5eea3c02361f21ad6ec6bec8f6a34d2be156296c42513da5533fb9d3af4038f7b02c9fd412b3b70f94bb65e569922e3e9f346c0b8ed2ad109e0d5f3b953e5411ddfa7ed7b0a60542e24b0510f3d1a96629ba0a58021eb623f3acd60ad8f79c4278dd5c9213020d6359fb392fec783bed11fc3484972230b510f0e63cadf00b0b65fa135c61fe0cbf3280ed4a745719d8945f1fa1864a507ff66b293b5c51bdcb35a527746d528fb38c049829187d0a677b124fa68b2d6ceb565e1321b86fc", @generic="4d721c8b383769a764937f17db741ac512ad5f84dc682af8edeff4a74f6090876cc97ba3b128eecb2eef11038b50a328d818c4776ad01890bf08440a9a"]}]}, 0x20c}, {&(0x7f0000000740)={0x1084, 0x10, 0x800, 0x70bd25, 0x25dfdbfd, "", [@typed={0x14, 0x5d, 0x0, 0x0, @ipv6=@loopback}, @typed={0x24, 0x18, 0x0, 0x0, @binary="17089d9abf899777e6a2d77f1e7d5724872ee91444fa4da34531d08dc7f9ae85"}, @typed={0xc, 0x128, 0x0, 0x0, @u64=0x9}, @typed={0x8, 0xe9, 0x0, 0x0, @pid}, @typed={0x1004, 0x25, 0x0, 0x0, @binary="4d25059a1a01986681c822bc09f098bb73c119783f81dbe8824a8fa9014888c3f670af704119f6cc1545d35174c987fb52d9051bf3b01e4c7a424f595280a47cc3a01b45f06cba1defe1ab7ce1d9f66c00f7e039720b0dfc9e1d53cb304618a32f26256dccd3da1d40e44d66bf9f3c413ebe925e39394fb082f7818758d0b7e58df5069d078aaf74adea049747759263eea91b63f86a6ddf0b3290685b897c6073700c5c1dd8fea2dd1d01602e4299d5f125cf0163fa2c1ebc6de3f042eb9bde2e479ebd7a738f35c4cdb587bef8e571e0edaf3c4705553180e1cf9bf0adf27fd2835864e25fad08d8cfaac58e6c60d720aa7903db9ccb41fcf8793e14f97c81867739fb11c90d583f0a35654979a5a00e8e43e542077da3c7d6f7466d5a00d5d60a23b518e501fcd3e489c97762f6173cbb973c450b3b26edeb7a730407c74e7f11aa25de3bfa907086945e67b54a837c73567ae62633dafebad4d38ab5813bdd9d80262b6d2471a1a61cb97dc4fcdebf0cd980c04319b0463b1d0db04131c411e535438233a2028650ace7688bec3de1806e44ea79a63aade453acfcdbf5aa3e63c3e075a5caf2bc745d88554b7d23afd40094c9e654a253050ccc8e1d549c6042c9f8da7cbea3898c8573b04b30a678bf6d39c0f68fd2cded90da3dc4a008dcb455a30765aba41fdf7b9324d7ba34ac8dbbd47bdbd1b022b78ce2b765342dd776059d3a79d86b317e199147b1933b0b1def209de8492bd6962fbc93338eea9a5b3288161294459c765bb8a982f8b16f258aec9efbaaf651667758f05a7e11163dcd36c75bccc3a6df491499ec6b5c9e86dbb37a495033bcf5fa1a4ab5de98a8810e717c7f8db4a22076ddc32271f4fcf615c0a80655a7bb304f1d020587f34e6e7c1b99b6102ef51a8d92d7373d849e170129f4a1801d45d5aaff670c07afddfdc55fb0a597951c58603585f247f5a11f9c35ec83a1680f103f8ecf6deead4ad441cd67a4049486a16ccb6ccb5e935bf3a002312a2c419a0b14dd436730bd9b7f5c74c36f6b3dbcd184d750d0278fb1f8184d0e7545b7d7af48d71ca0372ea5b8e9d8ad797dafce2418bb24513a2653a7786998669795d8146f5b7ce28fc183aada6a942d1e0df98c1dedb579a46ed84a97798e6c4415266c114c659b01cb35c88a6c4b4b0383c5fc2b06c551319ebc9040a9d2939cc8bc1f1e5d0e156ef1d4d782ed7d87a28f3ef15a4494ef55ec26eb665078fa23548b9a66e438c3f747ff20ab2ea6aa03298b45fac67ff91dc3aee6abc127d5ddfd1a6f8d91999f85bc95483ddf64ee31f0a3d93312b5b29083233478ac85e37f9550f12cebeb072d24932145c7a211f4420f2bcabc52cb4caeb92354021181554ad0a4ae1c0d9d90707990362236d42cb29147f27d6830a4123d0ac346ee69a91922107d80b3457a66f274d98ef2d867022b34379d229ac4e702a64db3bf011bd1b4daf0a6d878ee9b681c9f52e5560a20225e871fa7ac880881f1e39a602400ecfa89665bb2558181232ded9fdf3a1063f4a1ff36241fd0ad58f09c05471dfc263a52dee36cf97f0915c22e7b22abfece46d5b0650fa1517a0fc3f820d5c7b378256daeee7723d9f04858747d630d3fbfcae75210663075676d890c492a50a0548f039deb20fe0a33017c1c721df39883dd6484267e05fb244370fcb729b522eee69ef47ec7934a380531a5f42b139822b6c7cff00d5cfb291ba9e2edcdf8b8b0735e49ccc2dabe9a1c363feeb821b04a645d3686544a212e887aba2bc4b6f171873c465a8a70e0810f036a73eb86ffc67182460373fb3a1b858eaac511934177126ac3060b11480fbc4474b558c496a27221a55d72370b498e081ed17d1197a42b373ccb86d76a2487dd66a60acc61bcbe7f61b72bd9e533793b7afc1425f1e8133c791cd65b4e766642c9497fc5ccca4a530d0235a620d2b08e2bc6e2b940d7e2b6312d489f63d525ea2e8c4df4d4ac05435b63681b91b13ae0d203eaa722662cdc33b8ac68fbaa2ff7f657952dd46650bb09633c519cc7c44126ee05babc55a3fbc2933bedf7cffcc55d7a3fcc4460e040e2896d1d8070246c821eef4e6723bc1cab07b305a4fb4ac6d78898d12144f56e25c8c5df9e6c80c7f866c3961ecbbaf1eed026da06b776c5591e1af5a7954bbb4efb0c1af303b688aff0576b1c50b79e6e310ba020ce3dcb564dd606d650a3dc2716d157bf013211468536ad0e6646213eb3fb02a884dac52b54a5aac329ec1eb4a5fb2beaa82f4a245410b3e583cd3c0014fca640d7a752ca6dd47c271d1d4ada9bb79b42501cb4f5894b9b9681a43bfae67b234288ee50e0101decda623569d5d9dad0a7e4c608858fb04de1cf356c9f0db9ea9cb5df287fb23efa0b17082c1257d77c16539bccfd3e7591540eb82e9f765dbb5bb289f2bd489a0488032b5461d1348d670693f4e53341f2456ace8e2f59ac470dc8c999e9e52c2e0e320a0e2d1238296d1185561002d93825b147935945c6d0139fe99d3a155a5c1afc2f3f176dd521fe4927737d83d33ce48d333f39158f5653f8345b7cda03ef8dd217868e7b8e06ca3b771fb923458fb649aeeca8e9f08ff2db3872ffffa6e7f773c37da792fd1c3030fe94f1ef0b971b4b6632783f15dcadfaab99be10db107957061340730f37c1c5655117dc0115f33987e051e6acffaaff9491694e4de23d08a24a986c1e866dd1910d215e22b591069971355cb145f52fc22caf8f5b120a9f4f9723cae79e1d15f05a2494d730e31e548148e9b326f37d9565157d03415b9f62c631d41c7d8d0b40b5b99f02dadc6ab95d81a97098628cf1c2803228169e4aad29f298753ef948a89c9dd38d4e953b0f3ec6d97a33d49c097baea530ad9f766d51f695976398d8e10f8c3fcf16b2bdf8f1d2535accf1fa0e85edbc4f189dc4ce16c9a7330fc77864b67a6b296b1349efc28d4281320214024fa851f467d3acb8d3b56ec98ffc73a330d8c395c5b2b264b24dcd6d03d4a7f558b30be5b22c6707b42e855a08df63825a89157bf76031f574454b142f05fe90bc9eeee2d7669615aff7233e6c01dd10da16a0fb5b488429e1fd936fc60b4897bc62c708e6df7d244e1631879bed247ae681bf434196d8db50ab1f702775c567bbc7f811fe4d00682f1d0084f67610318bee61189ac300da5cacacf85f067b5b21c6d68cdd3d11d48f365c7a187c2895765131800ae6d4b64c37ab8211f3d870284edbce98669e6721969e5690876f6807b9233784bc58a50cc75350f21622cb9e582644571add1159c307e0c2d385eef4b92faf94826b52fe7075ee40d891114e9ca69a389e1d571cd4c66bee43876b1251828a1a181b4011197c3fdd2ee610dd463320afc1e2f17c1f23ef6b4dfb1a73dc8747c94060a741cabedbad0114bfff358d0508c977b784f0d352de6b7c1e49dfe47e99211483be99e7b33147394f0fd9bc87f3cf0c02d174fbb09ff70add88ac6d5e243245b09d8e2b0d349698bd38f75ccfbd989ceb1bc4bb5732eddb177f97c80e0438d5e1ba15c9ff2ff50ba19ad93b4356a2664a73c8cb33663097887d7fe88c1de3d78309ebe7a16f483f544406ac448aa9356bd9171e4aa027e43295bc260bb1b6ba1fb00a4ac4490d1dd47e84bd6593231e1ed6e1e4d1defe56d7843e5a0120c2d33a9303349f424fc4476f27e2454edf30954496e5b4e789802eae2229f43073ad5c1ef64ffa3f8b4d0219b8812f2991866c36a7b1dfd8b510f69d1d1549d26db9d3c3cf7aefe1665c81082f2b9fe58d0e7670e9fdcda32566cb396a5ff81852b93945c6720554f36dcbc79bd44b5145d9e8382d6277afa4eb6a28d36f8239a533be4f9ee1fe8447ed63909391208b8b3af9df22e834ec9718f3474e73532fee8e113066d8eb5c317e97b56e7aa0749b4d84d8dd5435000c05e6b73ad4978cdbcc47480737923eafc754ea4587ce6b9b08269a3a73838e93c226d50449eeda62858617f98846c0658e9ffa8ecd2559b21fea3db08f413b593fa8c3e09d572993fe7acb83da43d75a5a789e5bca383f32966271ef7d4c1c6be53b98a1a7ec2c23957cccf019f7a7c0f9a87d89f1c15bcedbcef2bc4ce081aff354bc3df005ec813693348af7a4f6fa81ae4ec3616d63b07e1397cab03dc36654503b38dfc02250e00ce7e6d73ea53437b467df05979636a9d50e0b326049db5e801c4925ef646d85eb02b46d40a596b5c9b05cf31f441b97afde728d4c4c8062603f18db5b60eb327bc39dfe963601561728d4537c1da7a6120204b8b756d950c8779aec13c1891e0c1dc090857d3a5c15c074c2c86698dd9dbb43b8d789498c500058a3ac00b04ba5b4c648a7192751fb1d32d40786774ea04ad350ba349b7c0e37c39cd25d2e93d9ff06b0be3646fad982ff23f4f98fdd724396b641746ee2d8f786e7e81b16b59f4e5990ec945bc0294145e63295618e6f94ea3ab91e8bf28e65a395a27716ee8cd93d8095ee9f8472d879a383290ba00d74185bd6a34cd205eda9c4812f37eb1216cb6af28c87aec4771d7e2d9f52e525a3cc9abfc72663e5d791d4350b4393c7cb5e0572eb96de78389d794f80ebd9bc92b49445f6b68c52e6c5a6c68cfdef5b7f59626fb2a8026b906bb8c0e86f62a364c16ba1bb9ccc8f8288cc6cced7e1ecd974090dabc3a517634ce63d589b7256398037402969a485e92ff6fc6d9a572d3b9d423c1f76ac1de491792f373aaa2bc89dfb343f321b9949918504f515897f0723beadecb233327cb163116eabb36bb6543cb0cb3c2cb1145797afcfc6d626797b28f5f03bc2cd1d7f555283233d50dac1dc1f805782952e8a34c0951ae8a6691c31fa1eac2bed1c33c8ee5596002b3c687ebcee2d47475276d1cc1627866b904f5c3330fdc47061b7e104b4c59a5d019317865e71c347661f0aaa5035cfb216bcc3427fc0cfe06fc8522f04e7118584183e8722dfea9bb629a365d0b77e0e4689598a407765da16483309770b1186262f9625d507f93da36f2cda11c97c498a3cff11d6f351346f882063a277acf0f80bd452e90a44c161893b4aaade86750aa6c1ddd6df3254e1dd4e3a18054a369fc9e98ebd81d7ec35bfecda1a6ab66ec8c63331369649446d9d0da3ab9c288b83c0db15863c7aabcc68c4249b8cb7d12b46d9baeb2752ed26d71ffe48ad6392693153b13e7c8cc326c12b5603cfc9f2172269f547d2529fc70291bb38e3bb3279b9d765969bbb62b0b905251de86f0032ec26b502e907ca47756acb5bf49ac9014690d0a112d5bcf748a0fc1a6a65e9b69280ace21e9b048d4c519f0bdfa4639c5075e1cd724f3483056bca45f273ebc304ad2c26a651c31559accc7170c4f227a260fcd6127d5283dbca6241631d2752902223ad84cc715b219ccb2cf7ee1f1624fcc63739d828c543bd2ba6353f784efe7dc0e9c7837d5fdad8bc35849b18299eecb7936231181a87a3db380760c98f02389c37c0ed6758ee80d5497d0e1b2f3367efcd0225cccc3a09db7a8c8801ff8dcd86d92b53def8b3a61dc00de3f151628603238fad3b1c1cd38d98a0da9e30ab6e7228eb9970549ff06391adfa918f1257a827626b9c304d0687d0c163c3563867c317cacd371313ca74d0d05326546e9df2f22f9b1c21de253991321a13452a3b800c64e0f33c4d9b232058ba40ab92b2010464b4ca8fc10de7baf4860d80ce4b948ef95f6c684cf3c061461eaf5341c09c596eeea339e626c2f4f7db554283d590fa9b83"}, @nested={0x24, 0xdd, 0x0, 0x1, [@typed={0x8, 0xa2, 0x0, 0x0, @fd=r0}, @typed={0x15, 0x82, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}]}, 0x1084}, {&(0x7f00000019c0)={0x1424, 0x17, 0x2, 0x70bd29, 0x25dfdbfc, "", [@generic="133f5ce256e48e186b57eb92e3d67968193a2cf400d7cf3041790bce471bc257c3438d3e71a7d8393ef1b223f84de683314deeff92769a9fd838660088a17f6c2d29881a441cc1f341dc6ee6f4c5844598db7afd982706b468a39ff3ea4ad1631488fad356047ccc217c45ca7627ffc75733e22f495b0f256e7210a8d146509c7af6c38561393519a3b182b78ec1dd9bd25477a3cf587b27", @generic="e88ec576b23c16f99f3075597fe305d7cca55f465552bc35edd73b9e4d5ce2b8b486a37fc83eb8ac46f197789352d0b11938a7ba84027eff3fad4ac1222bb0d37e0471e0171b48d59270a259b2744ff58a8de6debd54f67c78bb47fbd8121844e292a0cb7e8a95c97e149a4d8f3ff033d6975eac8d84cbd3bf146fff16bc4690f3f3ac657ca1b672ece1672fbdb017c389dc634bf6bfb31190643d731343196c3facfa4d7450b8271a53706e4d46f5f873e479f3e1f3304d18c78039dec6ce750ec093f530b30b046597087ca6a4cb344ced7bdae4bbbcb05d3ef335cad3a84c072e397d382e1a4badfd816f22f0e243e0fcf79bbf3a6ce8f2", @generic="143fdd42b4d1a20607d75360dc83c395c8c3ce8252a2ce7008653d1b5e53fb041a8ccac0cb52b2df15a1e74b2e4c43a4ea411c43ffd236f7c089da121b3494d15fb915c4a8f58de93e5a92789bc38b542602a9d8af3e242d4a2e", @typed={0x14, 0x7, 0x0, 0x0, @ipv6=@ipv4={'\x00', '\xff\xff', @multicast1}}, @nested={0x120, 0xc6, 0x0, 0x1, [@typed={0x8, 0x3a, 0x0, 0x0, @ipv4=@initdev={0xac, 0x1e, 0x1, 0x0}}, @typed={0x8, 0x75, 0x0, 0x0, @u32=0x401}, @generic="b6895074bb7a3afe9630c6f99260b1f06c93bc2c0722540649440807d7a24b97f13ff6d21106215013a3fd5699a435b16158e53afb0f6ae0ff043d6639be1fb09a7c68", @generic="2749fb2104172d73bb85cc455515bcaf861c1add239821475b56e3cb827b64c9e8c2862d0fe2709bf0adc4377a55e79ee35151f9038a77fd6c62c16c9c743284915f107162cfc72596d29436f2527a36488416958a5c68125d51c5144522d271b74c6073546aae75103ccc5369d329bc92d13548f43eb3ff88fb916774b5d66961283b0fbbb72316c3552ef5507be8157459bd55be305195f50488b1a1", @typed={0x8, 0x105, 0x0, 0x0, @ipv4=@multicast2}, @generic="fd03adb6efa816bd7ffbfd7b56324c6291c0d47943cb464325b3037bcd446f2e310f70b6"]}, @nested={0x10da, 0xd, 0x0, 0x1, [@generic="845dde559616927f6d889da26095d1f40f2712fc6c56ebc6a455054ab7", @typed={0x8, 0x82, 0x0, 0x0, @u32=0x3f}, @typed={0x4, 0x50}, @typed={0x8, 0x57, 0x0, 0x0, @u32}, @typed={0x4, 0xa5}, @typed={0x15, 0x10b, 0x0, 0x0, @str='/dev/virtual_nci\x00'}, @generic="ed4141bc4f65989a9acee3dac370708ba95194311416f97e58f4a032c20ce1e20cdda03f81f0af7eed5353f09df66fda0010f3adb90b890776cd0a4e3436e5be3ca67c868e0ac0d2afe1832180a82f908afcbf870998621b", @generic="035ae798dea74a23cca6ba07b89ba8e528f72a786258952651b2346be481a4d62a955679adbf5cb2c0a16f7a01a314546c02d732e756dfee390855b0d51d78a4f4725565deafbcecbb65ab1fa666fef767d6019ed11646ccc00c84e249142e19a026cfaad65b4d2180b754ab21aa2d8c6c89eb9b2de191e769a05af2d8dc62f9bc80a546c4dd13b1f23f75253431fe2e92c09a577c7ea70dd8d9a348ecf6e7ccbea1c0c67ff07ba9edf44e3b3e50cc631d175fbb1480363d63c066920b56209b292fd82d1f11c66b7155ac3740daafd76cb43e70aacbcf63ce4e63604b6f9b178a6bf3dd0ebe5261ecb144838673d88e77eaa8686a4f9bb7eee95ed1e5698275ba14d32366d90e12551165017478e4774d497dba0995f4aa962010697dc9d7743991ca5334f52acb6f46f1e76a6c7d353d9f4b2de20ee797c17c599391c5f94249bd7dc0475caf9c43f839a745cb210055959073d5ef602d12f1bf8db2c7e60093f5fbffa27ecd3575d6ca03dd468f854cec88606babd590edc6fa795f2525fad1dcb81cab9e980739e1622c9650d335f8f00751d87bfb980fd1940c3bc7b06bb73062faed4e73b635ebd5109ed8bc8c21c4ec776345cd488a59617224143305d9fdcb064e65b9109dc2389103987ee41468d1613987bffdd8bfe7929e7d2d799fb22a9bc8a400abb17c962497b5d6579c55db7979b8a60949ee31759f5013873d5f610056ab9057065381f86cb44c5c392a855eccb76d40276b0be45a6919163a5a2a60c7e735332c87cf8f56d01dc8122f01a55444e9d8857acff85d3776cb6973840969898b6e26cb3c2b4272481ac3f2ec0d4a431c37a6db5b543af703245b3d2e6f1a1f122cd4cdf978285aad0ef7be6e5e5738a41fe8a941ae84811dba1ba265351fc85d95325c40fc1d737992fca90fa0738b8cd2108f9238490ff4c7a84d56e6009568b6febaf03b723abf1742b8cf3311331b33c3b9cd1f25ccf25db2f06ec2d2b0664dffdf05831c51a1a08fc499f9e8e04c23b0f7b5106e65913a65baaf963e4d38b13c4de41cc39cacc0393909a61a6fb1d98455c4831b7f8295f95d1b5b6af46193fd97296fc66bb3319f9d1b75818ce9f3176de68d4f8c4f758b1157a4709222df95d3a34cbe094086cb78cad42b974b03828fe81afff1e8a733060d0f63a6a5d589cafd566b36c6eeb512dbee5805407a57859d2195c5e8032f961552e7585671490cddd0cc1e6f516dd9f838040fcaa94ce4367642ffcfe7a0c112a25381031465d990ab1a01b75a5966d04844065f1f118cf0d1c116b6c34f796957fc2228cfca4936d8036b0f4a5d39c12bd660c37aa8bd47a303583a520d3db871b2de6990098dbcaa66b5513e493b50da6ea8415414f95965a8424c727d48e1449780e8860513244b61429b8a2a6d0b881239e54882d05920e4730863f19b34f6f48d26e7c45a65625cd80ba359e5d47ea0373deced5f0a4e84f9718d2e6aa46adb3a9c87b650ffe925cd7063de5d72912c308ffa9f7d6641c9ec82e038f51493792b42fc8019b480d0d54985fc0f7f86c80aafefc4bb3f9da67fbe034d1713319ae95c47ced97e09d5e121212eb4f5e10a55987edc54f13739e534a9ab9839bf1f3a42f6311c748d465fc8e117a26766e73d59898e34d2eee291cd20cddd6f3ad1d07586eb4db733440a3179b53c87b39a185acfab259e70cf078f6b878ba8db4ae98d9748a13670feb16731117b9b10a957fd51394a3b412c836f6ce1af7070a2cb3d73b155f576d821d4491c8de854d02353f815974b6d5a560944d781530c1e7af0e302730ccb5aa38b4b1cc1cf452a0fc0e9a091798b2f4e35be94c8a8a72f07853ccbf3e226a0259a0051c7638c4a285e4d7f6b606e4eb707d51d8ef3b1df65e1cb48849a29de75adbb19429c2e022a05c751459cd0ff536b42e2186aea7a299ead87ee2a050859b05da1698c94be83dd74b4b466cd962557ad32865966732c352e527fa0b0407a6b0022766c64566a3ec4861937747f2e07fce824b2ca0d8d12e3538c8a3d25a87f8332a9cbc83f124bcc0b886f4cf82c0d6110cb8f6f0cf18aa3dc88f66a8e3f59fd4480c3adeb1492da7c7f177870cd9808f3231eef91ad6b622bcaf55643e3ddc8d3b1f8b57602eb4fffa127b78d8fe424c385e4e577f8f31d424c6f4fd86aa0667d98ac2a6de2f598a14a160f2495f69888d432812493a0ca775d3cdf40818c7ff9e5a2164f715c993625588a68ed971849330cfb0d32ae6360a3b7c2b0ea413242291706ebc4ad649ca1a57e9de7d99f574d3891bb2912c921b7a7f9a90bf7c428b6970b21b8158b0c913caeff7ae6dfc745d367cf5b0b93c4b39e8c9a0788360aa41785960326d4abb8c16a9dc25aa4f63609e5909966bede3a99084ac9b3040b68aa7ad5bd7cd79e932931bde3accd9a46db66b19c5919a68992452e1a4d3b1506ad99c4a56f66f8cdc02f50c3e904957e3fe3e72d0ac3f6f10c5dd838afafaa5a8a052ceb37fa97fdb9d8e71004aba543ecca85ec88d69db7d46bc1bf089a614a397ffed1c89f43e0bd3f5668d628f22574f713fa793d3ad8ba14c08bc7071424e4dcd821e5d914b29e6862afee6e2f695544e28de50dee2c9b31ac3d36b4a53188dd1f602ac3faac101218b92e19b396353680dba8b0d7a86d00222832f560269362a07c12b25b9a42d748dfc98b8cd05ec713a7f4e17fefb7ddd49a7a38a72bd11d8f2da98aff2521ed44a2dda6c5fe56f7eb78ce6f5945e2a0e04b7e74e692640c109560b8037a0b995bb50b951c765225a3736a03c5c999328a6e404d3627da9aea8e110ece0d57296a86c780c16e6d21c6f9d3c1e6997e00a9bcc40b7b02e3964e25f291dd2e658c8cdf3497d35c2597923fcf8a87223bde8166933754710b2cdcd0b9935f0c04cc97ce201966f0b945708875a428394b4c7faf661ba8b6d35f0ba5d4f6bd51502ae25893ba2b32147689448410a40a09e14b4077c11f271fcb46cddce4ae16e329c09061f9398916ffd6137e6d106afbe97db920abb260c1a218767449fbedb7919518c1bbca93a8b7f640468a0f987a8258d55ec2702ee7ad386f412b07d77920ca6cb2c36cb426de4f3e90092db6e8d2d786d520d0dc0441bc98dc6a10fec47f87b9ead5c11f232035225f826fab74b4a45279344cb38174ddb9fe7794017898ac714e0892098fa9f5f522d6c16c9af130f3bcd225d9f40324edcd7e7aa8031bc30460ba4285e8732fb5e3d0cb7736b05540e340ff7ed7fa540e166d1eb053f9c22e342e7d30d7bdc836ea5f1df25233e6e50a3520138a7eebe0ef6d72e8699b0f68b4fc99e899ba8d1669cbe7c9af296391083e15b5374162c03e1b0e5de3f0f250eb1d040cf2a4b8cfa4799c8281a4e8c92bc244730814d19eb2c14ecdaa525253e112d12757d49d92b4ae22b502781d5c20d5dfaf6f80d8db1a5b3dae00dd3146438e2cc573c93a1d7c3057a308255d002057f5f491463f7a018bb63ec9008b05a089b70512851c6449b4127efee6a8baaf4afc5581c49c48d5bb1db182bc936ddd0505ab7a941bd29b7e674cdf4c3327aaf53364271fd79f7f21fce359de940d9a82d17bd0df7cf25c27c45c49e3664178464ca791f809906de74bd710926561d5ffb651a37b6dbf12db36b7e3de67027b6a5557cac24cda44b86d741bbdb9ae65cca48d355077bea2d606f0e91357d47430fe7c7ee8df80a746fa4d553b089729dc9b3a4cc980314bc53e4a3c65578687408f1038ce7571630486a14f187308e291b3feb9799d4642dafb99192902b86e8ac8a487637a7288710073e0e4d0b1174b08f326f9c55debbf66468f4c1da90762b64b91b99717022a463bccc9b6bff95ae10cf46759bde811f6b22846b9acb152e978a82afb91ae44a2ca1094fe6a757adea40c7705e3d2e5549733da91f22f4e34ec59b3045cfda255e06da9b8e5eb17a95a43775d67a9add8e4b530656a2d6a5ed5bc8d78595d6d2688dcd09d51173a6670c7c71b6913dc92c08bb4e4de9522b338c581356c37a7ddaf509050b50b0e74e401c0b76fb0062407664cb076808aa59991593d453c1c7e090bac17b88eb6b5e6f0ae9dc35173be30fdaa7cecd8285a7235a1d8979e6a769143fe938bb4047167cd01055fd080f7d8d435fe7d3e03a53e7ca4598cb3b4c265352891b6b40872625f157ae43d703aaf9c4376d9f02c003f3488b90378fa1c4936829c09c70c6fa8f049883c909fadc7e0b492e2823ecde921c6050b7f36dea0047a9a9bbf837a414d46bcaf3fd0faa4422ef11f4054294d25e553203f0f9354756521d4805c14a7b694d0808c55455e06cba9d93ce10b3ba082f25ff35feca2e4413605f3e470715ad9f1c10d6feb246e848b4d9c138fe734640ff125308e157c1f22c2e49acd2ecce153d0ea9ef4bcfb83b8faab3e729c03d85f3791ace97ee6fb594bf8510576790d4c996c0179c7ed72ee90ffdb5e47dffa3e0baa9d4bcec2630c51ec8e7afeeda361eb73d5215a70ab010ef60a5b3b291adb4a183df513958ee46349e290da8ef362fc62baa4ac788dfd06e7b385f04a3f95222a38be90eba5db9fb58b674d09b5ba05c8b74459ae8d08234a561ca87665baeb401740fce053bd7574353c8662be0c48e67acff8ca07330825882801a4c7b845213f43da8913df086b48485046ea423fdfa55523a1930f283736a7a5614dabd3abd2901227bd8aa3331eee00a3fa3fda70e91114e89deee858dd5d199c566aa005af39fcb96def00c77ab79b5f3a48e2d679c8c2f00e115211188726e1d03d5173407f6166b468a4115599b613604008e46a909faf4b6587b55727c103f166a907928955988bc306ca654b250efb6351e73d5dd683a742a1749fc64080230f9892518471eac0600c1d6bee5dfce0898c503c455c5f460564edef7c2134cc389f19e1c4b05890458bbd9c9a769ee702e796e968bc3b6791b4c710fe9cefe24d596f3e4888584219bc28bbbeeaa97cc58b6c38f01531ded1fb781a73b82bbc4ac0fca9e005205bbee81197ea88171eacf6b65cefe01effeccf346c887dd900edbb39fbef8ca53e9cd734a7de98efe9893c72589971c7baa00ce100070fc6a905be104f8a168f8238ad6cd21eaa5f5906aa7000547b2aaa8c2f55f1c8dc81b25a87f16acd3d2711e392ae25c03dbab151c9a6b99e862c0f8da5851d903a3805c03368d10c74e0e650596c26c8829bf5b37418f2b7f7da24d85f19594d808a1eec88827653daf5a56a263c42cd9322d4f3fa37860d80f5a3ba0c4b9007f2355c181a383aba2b57b892fdb6f366f52f8066a5a092a9b123472044661be47970ddf3ad84e1b72b1f47b97d42420fec3f50cdc02f9199b83359510fc684f7e70485c6ddc9624b86a416411bde08f402ac8db8ad7ba15558921d62a59c968835c7ac40d72e74ecda85fa9ea15b9bb992339306b6727e559e19ad62e62d514f55fea9d5ca53b2547ac4842f14421e876aced3f1ec39e047541ec38b5e3595d4c272179e27eab4dadd8e0ab2f7789e410ec7c3f2392c478d6fd0373a1ac956ca218a6684fc9c37cbc644b7b3c0c3731e343477e0301d7ae5f512cfcea836c93f65d80595b92452ef7065c924a0dad1f5adbad200c2014b3843e807f77b3fc369af645f25673ab6e874345d47cc61e5461b537565fc84c9da6976bd404755deb569e696296d372ae7413de74deaaab241f375d32fdaebe1b73bd361e92e1d09838f0f4fe96011b403837994666e7ab5", @typed={0x8, 0x7b, 0x0, 0x0, @pid=0xffffffffffffffff}, @generic="cee42dd28e3f52aee8d7a91f41360c271dc994d6c09730284216ae66b25938a87480ed76f5bb06ade0"]}, @typed={0x15, 0x9f, 0x0, 0x0, @str='/dev/virtual_nci\x00'}]}, 0x1424}], 0x5, &(0x7f0000003140)=[@rights={{0x30, 0x1, 0x1, [0xffffffffffffffff, r1, r2, r3, r4, r5, r6, r7]}}, @rights={{0x2c, 0x1, 0x1, [r8, r9, r10, 0xffffffffffffffff, r11, r12, r13]}}, @rights={{0x1c, 0x1, 0x1, [r14, r15, r16]}}, @rights={{0x18, 0x1, 0x1, [r17, r18]}}], 0x98, 0x400c880}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r19 = socket$nl_generic(0x10, 0x3, 0x10) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) setsockopt$netlink_NETLINK_LISTEN_ALL_NSID(r19, 0x10e, 0x8, &(0x7f0000000040)=0x4, 0x4) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) 00:08:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 2: openat$nci(0xffffffffffffff9c, 0xffffffffffffffff, 0x2, 0x0) 00:08:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xf0, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xdf, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde9455686", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xf0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) socket$vsock_stream(0x28, 0x1, 0x0) (async) recvmsg(r0, 0x0, 0x0) (async) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) (async) 00:08:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x1800) 00:08:53 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x104, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf1, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaec", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x104}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 1: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) r1 = socket$netlink(0x10, 0x3, 0xa) sendmsg$netlink(r1, &(0x7f0000006240)={0x0, 0x0, &(0x7f00000061c0)=[{&(0x7f0000001380)={0x24, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x11, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="46aaf1acc3"]}]}, 0x24}], 0x1}, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:53 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 2: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) socket$netlink(0x10, 0x3, 0xa) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) [ 533.553645][T30554] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_tx_wq": -EINTR [ 533.836815][ T6182] Bluetooth: hci4: command 0x0409 tx timeout 00:08:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:53 executing program 3: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) recvfrom$rose(r0, 0x0, 0x0, 0x0, 0x0, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:53 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:53 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 2: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) recvmsg(r0, 0x0, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 1: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) 00:08:54 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) [ 534.325227][T30288] netdevsim netdevsim0 netdevsim0: renamed from eth0 00:08:54 executing program 1: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) [ 534.384020][T30683] __nla_validate_parse: 7 callbacks suppressed [ 534.384035][T30683] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. 00:08:54 executing program 4: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) ioctl$SYNC_IOC_MERGE(0xffffffffffffffff, 0xc0303e03, &(0x7f0000002e80)={"f97a91a2ae7660145af22dcf886054cb9aa6a0ff012460b529cc21d93d1583c2"}) openat$autofs(0xffffffffffffff9c, &(0x7f0000002ec0), 0x210040, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 2: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) openat$smackfs_onlycap(0xffffffffffffff9c, &(0x7f0000002e40), 0x2, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r0, &(0x7f0000000200)={0x0, 0x0, &(0x7f0000000180)=[{&(0x7f0000000000)={0x10}, 0x10}], 0x1, &(0x7f0000000240)=ANY=[], 0x18}, 0x0) getsockname$netlink(r0, &(0x7f0000000040), &(0x7f0000000080)=0xc) 00:08:54 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) [ 534.493417][T30288] netdevsim netdevsim0 netdevsim1: renamed from eth1 [ 534.539016][T30288] netdevsim netdevsim0 netdevsim2: renamed from eth2 00:08:54 executing program 1: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) syz_open_dev$ptys(0xc, 0x3, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x2000) [ 534.588078][T30703] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. 00:08:54 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 4: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000002e00)={{0x1}, &(0x7f0000001900), &(0x7f0000001940)='%pS \x00'}, 0x20) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 534.674786][T30288] netdevsim netdevsim0 netdevsim3: renamed from eth3 00:08:54 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r0, &(0x7f0000000200)={0x0, 0x0, &(0x7f0000000180)=[{&(0x7f0000000000)={0x10}, 0x10}], 0x1, &(0x7f0000000240)=ANY=[], 0x18}, 0x0) getsockname$netlink(r0, &(0x7f0000000040), &(0x7f0000000080)=0xc) 00:08:54 executing program 2: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) openat$autofs(0xffffffffffffff9c, &(0x7f00000018c0), 0x606040, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 1: r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 1: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) ioctl$TUNGETDEVNETNS(0xffffffffffffffff, 0x54e3, 0x0) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) 00:08:54 executing program 4: socket$vsock_stream(0x28, 0x1, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 534.844240][T30736] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:54 executing program 3: socket$vsock_stream(0x28, 0x1, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:54 executing program 3: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000001880), 0x4) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) [ 534.954411][ T137] hsr_slave_0: left promiscuous mode 00:08:54 executing program 3: socket$vsock_stream(0x28, 0x1, 0x0) getsockname(0xffffffffffffffff, &(0x7f0000000000)=@in={0x2, 0x0, @private}, &(0x7f0000000080)=0x80) 00:08:54 executing program 3: bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f00000006c0)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x6, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x5, 0x1}, 0x48) bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) [ 535.009493][ T137] hsr_slave_1: left promiscuous mode 00:08:55 executing program 3: r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, 0x0, &(0x7f0000000080)) 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={&(0x7f0000002f00)='./file0\x00', 0x0, 0x4010}, 0x18) [ 535.076731][ T137] batman_adv: batadv0: Interface deactivated: batadv_slave_0 [ 535.078659][ T137] batman_adv: batadv0: Removing interface: batadv_slave_0 00:08:55 executing program 3: r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, 0x0, 0x0) [ 535.085731][ T137] batman_adv: batadv0: Interface deactivated: batadv_slave_1 [ 535.096441][ T137] batman_adv: batadv0: Removing interface: batadv_slave_1 [ 535.103354][ T137] bridge_slave_1: left allmulticast mode [ 535.104827][ T137] bridge_slave_1: left promiscuous mode [ 535.108442][ T137] bridge0: port 2(bridge_slave_1) entered disabled state 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, 0x0, 0x0) [ 535.147678][ T137] bridge_slave_0: left allmulticast mode [ 535.150115][ T137] bridge_slave_0: left promiscuous mode [ 535.151535][ T137] bridge0: port 1(bridge_slave_0) entered disabled state 00:08:55 executing program 3: r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, 0x0, 0x0) 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, 0x0, 0x0) 00:08:55 executing program 3: r0 = socket$vsock_stream(0x28, 0x1, 0x0) getsockname(r0, 0x0, 0x0) 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, 0x0, 0x0) [ 535.256684][ T137] veth1_macvtap: left promiscuous mode [ 535.261147][ T137] veth0_macvtap: left promiscuous mode [ 535.262969][ T137] veth1_vlan: left promiscuous mode 00:08:55 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$netlink(r0, &(0x7f0000000200)={0x0, 0x0, &(0x7f0000000180)=[{&(0x7f0000000000)={0x10}, 0x10}], 0x1, &(0x7f0000000240)=ANY=[], 0x18}, 0x0) getsockname$netlink(r0, &(0x7f0000000040), &(0x7f0000000080)=0xc) [ 535.265513][ T137] veth0_vlan: left promiscuous mode 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={0x0, 0x0, 0x4010}, 0x18) 00:08:55 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000000100)=ANY=[@ANYBLOB="010040060100000018000000", @ANYRES32=r0, @ANYRES32=0x0, @ANYRES32=0x0, @ANYBLOB='./file0\x00']) recvmsg(r1, &(0x7f00000016c0)={&(0x7f0000000140)=@x25, 0x80, &(0x7f0000001600)=[{&(0x7f00000001c0)=""/142, 0x8e}, {&(0x7f0000000280)=""/151, 0x97}, {&(0x7f0000000340)=""/210, 0xd2}, {&(0x7f0000000440)=""/4096, 0x1000}, {&(0x7f0000001440)=""/114, 0x72}, {&(0x7f00000014c0)=""/184, 0xb8}, {&(0x7f0000001580)=""/117, 0x75}], 0x7, &(0x7f0000001680)=""/14, 0xe}, 0x9c822de0181430cf) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f00000000c0)='./file0\x00', r1) 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={0x0}, 0x18) 00:08:55 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async, rerun: 32) r0 = openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) (rerun: 32) ioctl$AUTOFS_DEV_IOCTL_REQUESTER(0xffffffffffffffff, 0xc018937b, &(0x7f0000000100)=ANY=[@ANYBLOB="010040060100000018000000", @ANYRES32=r0, @ANYRES32=0x0, @ANYRES32=0x0, @ANYBLOB='./file0\x00']) recvmsg(r1, &(0x7f00000016c0)={&(0x7f0000000140)=@x25, 0x80, &(0x7f0000001600)=[{&(0x7f00000001c0)=""/142, 0x8e}, {&(0x7f0000000280)=""/151, 0x97}, {&(0x7f0000000340)=""/210, 0xd2}, {&(0x7f0000000440)=""/4096, 0x1000}, {&(0x7f0000001440)=""/114, 0x72}, {&(0x7f00000014c0)=""/184, 0xb8}, {&(0x7f0000001580)=""/117, 0x75}], 0x7, &(0x7f0000001680)=""/14, 0xe}, 0x9c822de0181430cf) (async) fsconfig$FSCONFIG_SET_PATH_EMPTY(0xffffffffffffffff, 0x4, &(0x7f0000000040)='/dev/virtual_nci\x00', &(0x7f00000000c0)='./file0\x00', r1) [ 535.586063][ T137] team0 (unregistering): Port device team_slave_1 removed [ 535.595036][ T137] team0 (unregistering): Port device team_slave_0 removed [ 535.602691][ T137] bond0 (unregistering): (slave bond_slave_1): Releasing backup interface [ 535.642019][ T137] bond0 (unregistering): (slave bond_slave_0): Releasing backup interface [ 535.662432][T30826] workqueue: Failed to create a rescuer kthread for wq "nfc12_nci_rx_wq": -EINTR [ 535.675972][T30822] workqueue: Failed to create a rescuer kthread for wq "nfc12_nci_cmd_wq": -EINTR [ 535.801369][ T137] bond0 (unregistering): Released all slaves 00:08:55 executing program 3: bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={0x0}, 0x18) 00:08:55 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) [ 535.901655][T30766] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. [ 535.917056][ T6182] Bluetooth: hci4: command 0x041b tx timeout 00:08:55 executing program 2: bpf$OBJ_GET_PROG(0x7, &(0x7f0000002f40)=@o_path={0x0}, 0x18) 00:08:55 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:55 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x100, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 535.952608][T30288] 8021q: adding VLAN 0 to HW filter on device bond0 00:08:55 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x110, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfe, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c4", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x110}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 535.998106][T30288] 8021q: adding VLAN 0 to HW filter on device team0 00:08:55 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) [ 536.002861][ T6266] bridge0: port 1(bridge_slave_0) entered blocking state [ 536.004838][ T6266] bridge0: port 1(bridge_slave_0) entered forwarding state [ 536.014351][ T6167] bridge0: port 2(bridge_slave_1) entered blocking state [ 536.016328][ T6167] bridge0: port 2(bridge_slave_1) entered forwarding state [ 536.024340][T30874] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.1'. 00:08:56 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x0) (async) [ 536.045159][T30288] hsr0: Slave A (hsr_slave_0) is not up; please bring it up to get a fully working HSR network 00:08:56 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 536.061063][T30288] hsr0: Slave B (hsr_slave_1) is not up; please bring it up to get a fully working HSR network 00:08:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:56 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x2500) [ 536.201690][T30902] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 536.214695][T30288] 8021q: adding VLAN 0 to HW filter on device batadv0 00:08:56 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x4000000) [ 536.296619][T30288] veth0_vlan: entered promiscuous mode [ 536.301357][T30288] veth1_vlan: entered promiscuous mode [ 536.304622][T30887] workqueue: Failed to create a rescuer kthread for wq "nfc5_nci_cmd_wq": -EINTR 00:08:56 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x3f00) [ 536.306122][T30898] workqueue: Failed to create a rescuer kthread for wq "nfc5_nci_cmd_wq": -EINTR [ 536.339185][T30926] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.1'. 00:08:56 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x101, 0x0, 0x0, 0x1, [@typed={0x8, 0x21, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:56 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, 0x0, 0x0, 0x80000) [ 536.346618][T30921] workqueue: Failed to create a rescuer kthread for wq "nfc2_nci_cmd_wq": -EINTR [ 536.399547][T30288] veth0_macvtap: entered promiscuous mode 00:08:56 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) 00:08:56 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 536.462067][T30288] veth1_macvtap: entered promiscuous mode 00:08:56 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x4000000) [ 536.499001][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3e) already exists on: batadv_slave_0 [ 536.501736][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.504173][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3e) already exists on: batadv_slave_0 00:08:56 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 536.554808][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.570207][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3e) already exists on: batadv_slave_0 [ 536.583374][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.590833][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3e) already exists on: batadv_slave_0 [ 536.593889][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.602381][T30288] batman_adv: batadv0: Interface activated: batadv_slave_0 [ 536.604837][T30949] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.2'. [ 536.613518][T30951] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.4'. [ 536.619630][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3f) already exists on: batadv_slave_1 [ 536.622290][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.624708][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3f) already exists on: batadv_slave_1 [ 536.628039][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.630549][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3f) already exists on: batadv_slave_1 [ 536.633153][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.635961][T30288] batman_adv: The newly added mac address (aa:aa:aa:aa:aa:3f) already exists on: batadv_slave_1 00:08:56 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x3f00) [ 536.638943][T30288] batman_adv: It is strongly recommended to keep mac addresses unique to avoid problems! [ 536.642615][T30288] batman_adv: batadv0: Interface activated: batadv_slave_1 [ 536.656692][T30288] netdevsim netdevsim0 netdevsim0: set [1, 0] type 2 family 0 port 6081 - 0 [ 536.659117][T30288] netdevsim netdevsim0 netdevsim1: set [1, 0] type 2 family 0 port 6081 - 0 [ 536.661244][T30288] netdevsim netdevsim0 netdevsim2: set [1, 0] type 2 family 0 port 6081 - 0 [ 536.663467][T30288] netdevsim netdevsim0 netdevsim3: set [1, 0] type 2 family 0 port 6081 - 0 00:08:56 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) 00:08:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) [ 536.841001][T30982] netlink: 248 bytes leftover after parsing attributes in process `syz-executor.3'. 00:08:56 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10c, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xfa, 0x0, 0x0, 0x1, [@generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x10c}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 536.866372][ T6266] wlan0: Created IBSS using preconfigured BSSID 50:50:50:50:50:50 [ 536.868315][ T6266] wlan0: Creating new IBSS network, BSSID 50:50:50:50:50:50 00:08:56 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={&(0x7f0000000040), 0xc, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4, 0xd3}, @typed={0x14, 0x100, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0xf9, 0x0, 0x0, @u64}, @generic, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) [ 536.905577][ T666] wlan1: Created IBSS using preconfigured BSSID 50:50:50:50:50:50 [ 536.926399][ T666] wlan1: Creating new IBSS network, BSSID 50:50:50:50:50:50 00:08:56 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:56 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x4000) 00:08:56 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) accept4$rose(r0, 0x0, 0x0, 0x80000) 00:08:57 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) [ 537.104369][T30986] BUG: MAX_LOCKDEP_KEYS too low! [ 537.105705][T30986] turning off the locking correctness validator. [ 537.107375][T30986] CPU: 0 PID: 30986 Comm: syz-executor.4 Not tainted 6.8.0-rc7-syzkaller-g707081b61156 #0 [ 537.110008][T30986] Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 02/29/2024 [ 537.112575][T30986] Call trace: [ 537.113361][T30986] dump_backtrace+0x1b8/0x1e4 [ 537.114638][T30986] show_stack+0x2c/0x3c [ 537.115726][T30986] dump_stack_lvl+0xd0/0x124 [ 537.116945][T30986] dump_stack+0x1c/0x28 [ 537.118008][T30986] register_lock_class+0x568/0x6ac [ 537.119336][T30986] __lock_acquire+0x184/0x763c [ 537.120560][T30986] lock_acquire+0x23c/0x71c [ 537.121664][T30986] __flush_workqueue+0x14c/0x11c4 [ 537.123006][T30986] drain_workqueue+0xb8/0x32c [ 537.124254][T30986] destroy_workqueue+0xb8/0xdc4 [ 537.125502][T30986] nci_unregister_device+0xac/0x21c [ 537.126767][T30986] virtual_ncidev_close+0x5c/0xa0 [ 537.128029][T30986] __fput+0x30c/0x738 [ 537.129062][T30986] __fput_sync+0x60/0x9c [ 537.130132][T30986] __arm64_sys_close+0x150/0x1e0 [ 537.131419][T30986] invoke_syscall+0x98/0x2b8 [ 537.132585][T30986] el0_svc_common+0x130/0x23c [ 537.133773][T30986] do_el0_svc+0x48/0x58 [ 537.134918][T30986] el0_svc+0x54/0x168 [ 537.135972][T30986] el0t_64_sync_handler+0x84/0xfc [ 537.137263][T30986] el0t_64_sync+0x190/0x194 00:08:57 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 0: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:57 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:57 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:57 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 2: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 4: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x4000000) 00:08:57 executing program 4: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) 00:08:57 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:57 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) 00:08:57 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) 00:08:57 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:57 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:57 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: syz_init_net_socket$rose(0xb, 0x5, 0x0) 00:08:57 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, &(0x7f0000000100)) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:57 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:57 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 0: socket$netlink(0x10, 0x3, 0x0) 00:08:57 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 4: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000003000), 0x2, 0x4000) 00:08:57 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:57 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:57 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:57 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:57 executing program 1: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 4: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:57 executing program 1: socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:57 executing program 0: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 0: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:57 executing program 4: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:57 executing program 2: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) [ 537.996379][ T6182] Bluetooth: hci4: command 0x040f tx timeout 00:08:58 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:58 executing program 4: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 4: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:58 executing program 2: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 3: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 2: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 1: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) accept4$rose(r1, 0x0, 0x0, 0x800) 00:08:58 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:58 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 1: openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) 00:08:58 executing program 3: openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x800) 00:08:58 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x0, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 4: openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:08:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 2: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x800) 00:08:58 executing program 2: openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:08:58 executing program 1: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 2: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, &(0x7f0000000100)) 00:08:58 executing program 4: openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:08:58 executing program 3: accept4$rose(0xffffffffffffffff, 0x0, 0x0, 0x0) 00:08:58 executing program 0: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) r2 = accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) getsockopt$rose(r1, 0x104, 0x7, &(0x7f0000000300), &(0x7f00000000c0)=0x4) recvmsg(r2, &(0x7f0000000340)={&(0x7f0000000100)=@nfc, 0x80, &(0x7f00000002c0)=[{&(0x7f0000000180)=""/94, 0x5e}, {&(0x7f0000000200)=""/140, 0x8c}], 0x2, &(0x7f0000000300)}, 0x40000100) ioctl$SIOCRSGCAUSE(r0, 0x89e0, &(0x7f0000000400)) 00:08:58 executing program 1: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 2: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) r1 = accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) accept4$rose(r1, 0x0, 0x0, 0x800) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:58 executing program 2: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) accept4$rose(r0, &(0x7f0000000000)=@full={0xb, @dev, @default, 0x0, [@default, @default, @null, @default, @rose]}, &(0x7f0000000040)=0x40, 0x80000) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 1: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:58 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 3: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 2: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:58 executing program 0: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 0: ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 3: socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 1: socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 3: r0 = openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 4: socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 0: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:58 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: r0 = openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 2: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, 0x0, 0x40) 00:08:58 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:58 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 3: r0 = openat$sw_sync_info(0xffffffffffffff9c, 0x0, 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:08:58 executing program 3: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, 0x0, 0x40) 00:08:58 executing program 4: openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x4000000) 00:08:58 executing program 4: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 3: openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 1: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, 0x0, 0x40) 00:08:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) 00:08:58 executing program 1: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 3: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 0: openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(0xffffffffffffffff, 0x890c, &(0x7f0000000380)={@remote={0xcc, 0xcc, 0xcc, 0xcc, 0x3}, 0xf7d2, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @rose={'rose', 0x0}, 0x7, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @null, @null, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}]}) 00:08:58 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1}, 0x0) 00:08:58 executing program 4: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:58 executing program 1: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:58 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 4: r0 = openat$sw_sync_info(0xffffffffffffff9c, &(0x7f0000000080), 0x0, 0x0) ioctl$sock_rose_SIOCDELRT(r0, 0x890c, 0x0) 00:08:58 executing program 2: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, 0x0}, 0x40) 00:08:58 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:58 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:58 executing program 4: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:59 executing program 2: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, 0x0}, 0x0) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 0: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:59 executing program 1: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x0) 00:08:59 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:59 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 1: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:59 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 1: socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:59 executing program 3: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 0: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, &(0x7f00000000c0), &(0x7f0000000100)=0x4) 00:08:59 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 0: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:59 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:59 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:08:59 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 0: getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:59 executing program 1: getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(0xffffffffffffffff, 0x104, 0x4, 0x0, 0x0) 00:08:59 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) [ 539.593749][T22946] Bluetooth: hci0: unexpected cc 0x0c03 length: 249 > 1 [ 539.595853][T22946] Bluetooth: hci0: unexpected cc 0x1003 length: 249 > 9 [ 539.597845][T22946] Bluetooth: hci0: unexpected cc 0x1001 length: 249 > 9 [ 539.599973][T22946] Bluetooth: hci0: unexpected cc 0x0c23 length: 249 > 4 [ 539.601978][T22946] Bluetooth: hci0: unexpected cc 0x0c25 length: 249 > 3 [ 539.603918][T22946] Bluetooth: hci0: unexpected cc 0x0c38 length: 249 > 2 [ 539.651943][T31658] chnl_net:caif_netlink_parms(): no params data found [ 539.690281][T31658] bridge0: port 1(bridge_slave_0) entered blocking state [ 539.692539][T31658] bridge0: port 1(bridge_slave_0) entered disabled state [ 539.694781][T31658] bridge_slave_0: entered allmulticast mode [ 539.696688][T31658] bridge_slave_0: entered promiscuous mode [ 539.700480][T31658] bridge0: port 2(bridge_slave_1) entered blocking state [ 539.702306][T31658] bridge0: port 2(bridge_slave_1) entered disabled state [ 539.704244][T31658] bridge_slave_1: entered allmulticast mode [ 539.706093][T31658] bridge_slave_1: entered promiscuous mode [ 539.721320][T31658] bond0: (slave bond_slave_0): Enslaving as an active interface with an up link [ 539.724562][T31658] bond0: (slave bond_slave_1): Enslaving as an active interface with an up link [ 539.733637][T31658] team0: Port device team_slave_0 added [ 539.735868][T31658] team0: Port device team_slave_1 added [ 539.766763][T31658] batman_adv: batadv0: Adding interface: batadv_slave_0 [ 539.768550][T31658] batman_adv: batadv0: The MTU of interface batadv_slave_0 is too small (1500) to handle the transport of batman-adv packets. Packets going over this interface will be fragmented on layer2 which could impact the performance. Setting the MTU to 1560 would solve the problem. [ 539.775144][T31658] batman_adv: batadv0: Not using interface batadv_slave_0 (retrying later): interface not active [ 539.783923][T31658] batman_adv: batadv0: Adding interface: batadv_slave_1 [ 539.785722][T31658] batman_adv: batadv0: The MTU of interface batadv_slave_1 is too small (1500) to handle the transport of batman-adv packets. Packets going over this interface will be fragmented on layer2 which could impact the performance. Setting the MTU to 1560 would solve the problem. [ 539.792407][T31658] batman_adv: batadv0: Not using interface batadv_slave_1 (retrying later): interface not active [ 539.848324][T31658] hsr_slave_0: entered promiscuous mode 00:08:59 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:08:59 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) [ 539.893187][T31658] hsr_slave_1: entered promiscuous mode 00:08:59 executing program 2: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, 0x0}, 0x0) 00:08:59 executing program 3: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 539.937719][T31658] debugfs: Directory 'hsr0' with parent 'hsr' already present! [ 539.956767][T31658] Cannot create hsr debugfs directory 00:08:59 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000080), 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:08:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:08:59 executing program 4: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) [ 540.076499][ T6182] Bluetooth: hci4: command 0x0419 tx timeout 00:09:00 executing program 4: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f00000000c0), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) [ 540.188223][ T13] netdevsim netdevsim1 netdevsim3 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) [ 540.257250][ T13] netdevsim netdevsim1 netdevsim2 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:09:00 executing program 4: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) [ 540.316043][ T13] netdevsim netdevsim1 netdevsim1 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: openat$nci(0xffffffffffffff9c, &(0x7f0000001980), 0x2, 0x0) r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) [ 540.360268][ T13] netdevsim netdevsim1 netdevsim0 (unregistering): unset [1, 0] type 2 family 0 port 6081 - 0 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: r0 = syz_init_net_socket$rose(0xb, 0x5, 0x0) getsockopt$rose(r0, 0x104, 0x4, 0x0, 0x0) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:00 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10, 0x21, 0x1}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 4: r0 = socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(r0, &(0x7f00000002c0)={0x0, 0x0, 0x0}, 0x0) 00:09:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 2: socket$netlink(0x10, 0x3, 0x0) r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r0}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x108, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xf6, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be"]}]}, 0x108}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680), 0x0, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{0x0}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 4: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0xe4, 0x21, 0x1, 0x0, 0x0, "", [@nested={0xd2, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic, @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0xe4}], 0x1, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r1}}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 2: sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x1, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 3: socket$inet6_udp(0xa, 0x2, 0x0) recvmsg(0xffffffffffffffff, &(0x7f00000002c0)={&(0x7f0000000100)=@pppoe={0x18, 0x0, {0x0, @dev}}, 0x80, &(0x7f0000000200)=[{&(0x7f0000000180)=""/72, 0x48}, {&(0x7f0000000040)=""/8, 0x8}], 0x2, &(0x7f0000000240)=""/67, 0x43}, 0x40) [ 541.676633][ T6182] Bluetooth: hci0: command 0x0409 tx timeout 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x0, 0x0, 0x0, 0x0, "", [@nested={0x102, 0x0, 0x0, 0x1, [@typed={0x8, 0x0, 0x0, 0x0, @u32=0x3f}, @generic="aa1beea98e9b043ad370502a17dae9f93a1beb96b8b1db4296f2077f3b0bba480d0c9c87035dac0ef15ae56223f80af3708414f8b04a83f9fa31275596a0ad056eadb0326ec62418d919129e32c4145451e3d0f213352bda0225142761949b686e03ce69f38ecde94556865ce295470494d61e810c22305531ea75aaecd6cca66058faa8fcc2560833c49050bb99", @typed={0x8, 0x0, 0x0, 0x0, @u32}, @typed={0x4}, @typed={0x14, 0x0, 0x0, 0x0, @ipv6=@mcast2}, @typed={0xc, 0x0, 0x0, 0x0, @u64}, @generic="b2f360cd29302f6cf25b189403f54aa4af4a24845d3ab79d8fa44752e273c0d41beb1eb501769f70171c9625af5370be", @typed={0xc, 0x0, 0x0, 0x0, @u64}]}]}, 0x114}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 0: sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) [ 541.705432][ T13] hsr_slave_0: left promiscuous mode 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780), 0x0, 0x80}, 0x4000000) 00:09:01 executing program 2: socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)=[@cred={{0x1c}}], 0x20, 0x80}, 0x4000000) 00:09:01 executing program 3: sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) [ 541.736684][ T13] hsr_slave_1: left promiscuous mode 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)}, 0x4000000) 00:09:01 executing program 2: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, 0x0, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 3: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)}, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 3: syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) [ 541.806721][ T13] batman_adv: batadv0: Interface deactivated: batadv_slave_0 00:09:01 executing program 2: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, 0x0, 0x4000000) [ 541.812295][ T13] batman_adv: batadv0: Removing interface: batadv_slave_0 [ 541.815404][ T13] batman_adv: batadv0: Interface deactivated: batadv_slave_1 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)}, 0x4000000) 00:09:01 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) [ 541.828308][ T13] batman_adv: batadv0: Removing interface: batadv_slave_1 [ 541.835008][ T13] bridge_slave_1: left allmulticast mode [ 541.839920][ T13] bridge_slave_1: left promiscuous mode [ 541.841465][ T13] bridge0: port 2(bridge_slave_1) entered disabled state 00:09:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, 0x0, 0x0, 0x80}, 0x4000000) 00:09:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(r0, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x10}, 0x10}], 0x1, &(0x7f0000000780)}, 0x0) 00:09:01 executing program 0: r0 = syz_clone(0x0, 0x0, 0x0, 0x0, 0x0, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000007c0)={0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000080)={0x114, 0x21, 0x1, 0x0, 0x0, "", [@nested={0x102 VM DIAGNOSIS: Warning: Permanently added '10.128.10.47' (ED25519) to the list of known hosts. lock-classes: 8177 [max: 8192] direct dependencies: 32758 [max: 131072] indirect dependencies: 190476 all direct dependencies: 2578630 dependency chains: 59511 [max: 65536] dependency chain hlocks used: 243564 [max: 327680] dependency chain hlocks lost: 0 in-hardirq chains: 120 in-softirq chains: 1509 in-process chains: 57879 stack-trace entries: 282770 [max: 1048576] number of stack traces: 12807 number of stack hash chains: 8821 combined max dependencies:hardirq-safe locks: 57 hardirq-unsafe locks: 7577 softirq-safe locks: 267 softirq-unsafe locks: 7291 irq-safe locks: 279 irq-unsafe locks: 7577 hardirq-read-safe locks: 4 hardirq-read-unsafe locks: 167 softirq-read-safe locks: 17 softirq-read-unsafe locks: 156 irq-read-safe locks: 17 irq-read-unsafe locks: 167 uncategorized locks: 325 unused locks: 1 max locking depth: 20 max bfs queue depth: 1790 max lock class index: 8191 debug_locks: 0 zapped classes: 10221 zapped lock chains: 55335 large chain blocks: 1 all lock classes: FD: 1 BD: 219 -...: (console_sem).lock FD: 167 BD: 5 +.+.: console_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->kbd_event_lock ->(console_sem).lock ->console_owner_lock ->fs_reclaim ->&x->wait#9 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#11 ->&fb_info->lock ->vt_event_lock ->&base->lock ->subsys mutex#6 ->&helper->lock ->&helper->damage_lock ->&rq->__lock ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 1 ....: console_srcu FD: 32 BD: 1 +.+.: fill_pool_map-wait-type-override ->pool_lock#2 ->pool_lock ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->&n->list_lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 2 BD: 3660 -.-.: &obj_hash[i].lock ->pool_lock FD: 1 BD: 3659 -.-.: pool_lock FD: 858 BD: 15 +.+.: cgroup_mutex ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&obj_hash[i].lock ->cgroup_file_kn_lock ->css_set_lock ->blkcg_pol_mutex ->percpu_counters_lock ->shrinker_mutex ->&base->lock ->devcgroup_mutex ->cpu_hotplug_lock ->fs_reclaim ->&x->wait#2 ->&rq->__lock ->cgroup_mutex.wait_lock ->&n->list_lock ->cgroup_rstat_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->cpuset_mutex ->&dom->lock ->batched_entropy_u32.lock ->cgroup_idr_lock ->task_group_lock ->(wq_completion)cpuset_migrate_mm ->&wq->mutex ->&____s->seqcount#2 ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->remove_cache_srcu FD: 49 BD: 1 +.+.: fixmap_lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 FD: 228 BD: 86 ++++: cpu_hotplug_lock ->jump_label_mutex ->cpuhp_state_mutex ->wq_pool_mutex ->freezer_mutex ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&ACCESS_PRIVATE(rtpcp, lock) ->smpboot_threads_lock ->&obj_hash[i].lock ->&pool->lock ->&rq->__lock ->&x->wait#5 ->mem_hotplug_lock ->mem_hotplug_lock.waiters.lock ->mem_hotplug_lock.rss.gp_wait.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->rcu_node_0 ->&swhash->hlist_mutex ->pmus_lock ->pcp_batch_high_lock ->&xa->xa_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->kthread_create_lock ->&p->pi_lock ->&x->wait ->wq_pool_attach_mutex ->pcpu_alloc_mutex ->sparse_irq_lock ->cpu_hotplug_lock.waiters.lock ->&x->wait#6 ->cpuhp_state-up ->stop_cpus_mutex ->&wq->mutex ->flush_lock ->xps_map_mutex ->css_set_lock ->cpuset_mutex ->cgroup_threadgroup_rwsem ->cgroup_threadgroup_rwsem.waiters.lock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->wq_pool_mutex.wait_lock ->&list->lock#12 FD: 28 BD: 91 +.+.: jump_label_mutex ->patch_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 96 BD: 10 +.+.: console_mutex ->&port_lock_key ->syslog_lock ->(console_sem).lock ->&rq->__lock ->&root->kernfs_rwsem ->kernfs_notify_lock FD: 1 BD: 109 ..-.: input_pool.lock FD: 1 BD: 3643 ..-.: base_crng.lock FD: 1 BD: 92 ....: patch_lock FD: 1 BD: 1 ....: rcu_read_lock FD: 1 BD: 1 ....: crng_init_wait.lock FD: 1 BD: 1 ....: early_pfn_lock FD: 1 BD: 13 ....: devtree_lock FD: 1 BD: 1 ....: rcu_read_lock_sched FD: 10 BD: 91 ++++: resource_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 1 BD: 1 ....: restart_handler_list.lock FD: 1 BD: 1 +.+.: system_transition_mutex FD: 3 BD: 779 ..-.: pcpu_lock ->stock_lock FD: 1 BD: 1 ....: debug_hook_lock FD: 2 BD: 1 ....: zonelist_update_seq ->zonelist_update_seq.seqcount FD: 1 BD: 2 ....: zonelist_update_seq.seqcount FD: 161 BD: 87 +.+.: cpuhp_state_mutex ->cpuhp_state-down ->cpuhp_state-up ->resource_lock ->pool_lock#2 ->(console_sem).lock ->clockevents_lock ->&irq_desc_lock_class ->&p->pi_lock ->&x->wait#6 ->&rq->__lock ->fs_reclaim ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->crypto_alg_sem ->scomp_lock FD: 2 BD: 1026 ..-.: &zone->lock ->&____s->seqcount FD: 1 BD: 3622 .-.-: &____s->seqcount FD: 26 BD: 95 +.+.: &pcp->lock ->&zone->lock FD: 1 BD: 3686 -.-.: pool_lock#2 FD: 57 BD: 162 +.+.: pcpu_alloc_mutex ->pcpu_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->&c->lock ->&____s->seqcount#2 ->pcpu_alloc_mutex.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->key ->percpu_counters_lock ->&n->list_lock FD: 6 BD: 3608 -.-.: &n->list_lock ->&c->lock FD: 5 BD: 3634 -.-.: &c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 101 BD: 59 +.+.: slab_mutex ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->fs_reclaim ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->lock ->&root->kernfs_rwsem ->&k->list_lock ->&obj_hash[i].lock FD: 65 BD: 3 +.+.: trace_types_lock ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 FD: 1 BD: 2 ....: panic_notifier_list.lock FD: 1 BD: 1 ....: die_chain.lock FD: 51 BD: 4 +.+.: trace_event_sem ->trace_event_ida.xa_lock ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->eventfs_mutex FD: 3 BD: 300 ..-.: batched_entropy_u32.lock ->crngs.lock FD: 2 BD: 3642 ..-.: crngs.lock ->base_crng.lock FD: 17 BD: 413 +.+.: sysctl_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 23 BD: 3514 -.-.: &rq->__lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->&rt_b->rt_runtime_lock ->&cp->lock ->&rt_rq->rt_runtime_lock ->&____s->seqcount ->pool_lock#2 ->cpu_asid_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 3515 ....: &cfs_b->lock FD: 24 BD: 1 ....: init_task.pi_lock ->&rq->__lock FD: 1 BD: 1 ....: init_task.vtime_seqcount FD: 61 BD: 90 +.+.: wq_pool_mutex ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&wq->mutex ->&obj_hash[i].lock ->fs_reclaim ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->wq_pool_attach_mutex ->&pool->lock ->&xa->xa_lock ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock ->wq_pool_mutex.wait_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 36 BD: 117 +.+.: &wq->mutex ->&pool->lock ->&x->wait#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock FD: 29 BD: 654 -.-.: &pool->lock ->&obj_hash[i].lock ->&p->pi_lock ->pool_lock#2 ->(worker)->lock ->wq_mayday_lock ->&base->lock ->&x->wait#10 FD: 49 BD: 57 +.+.: shrinker_mutex ->pool_lock#2 ->fs_reclaim FD: 1 BD: 811 -.-.: rcu_node_0 FD: 10 BD: 48 -.-.: rcu_state.barrier_lock ->rcu_node_0 ->&obj_hash[i].lock FD: 30 BD: 3 ....: &rnp->exp_poll_lock FD: 9 BD: 5 ....: trace_event_ida.xa_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 FD: 1 BD: 1 ....: trigger_cmd_mutex FD: 16 BD: 253 +.+.: free_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->init_mm.page_table_lock ->quarantine_lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 253 +.+.: vmap_area_lock FD: 241 BD: 1 ....: acpi_probe_mutex ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock ->resource_lock ->&c->lock ->cpu_hotplug_lock ->(console_sem).lock ->irq_domain_mutex ->pcpu_alloc_mutex ->&domain->mutex ->&desc->request_mutex ->&irq_desc_lock_class ->cpu_pm_notifier.lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->iort_msi_chip_lock ->its_lock ->efi_mem_reserve_persistent_lock ->lpi_range_lock ->syscore_ops_lock ->clocksource_mutex FD: 13 BD: 254 +.+.: init_mm.page_table_lock ->&obj_hash[i].lock ->&base->lock FD: 49 BD: 5 +.+.: irq_domain_mutex ->pool_lock#2 ->fs_reclaim FD: 117 BD: 7 +.+.: &domain->mutex ->sparse_irq_lock ->pool_lock#2 ->&irq_desc_lock_class ->fs_reclaim ->&obj_hash[i].lock ->&its->dev_alloc_lock ->&c->lock ->&____s->seqcount FD: 114 BD: 93 +.+.: sparse_irq_lock ->&____s->seqcount ->pool_lock#2 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&c->lock ->(cpu_running).wait.lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->&x->wait#6 ->&p->pi_lock ->&irq_desc_lock_class ->fs_reclaim ->lock ->&root->kernfs_rwsem ->batched_entropy_u8.lock ->kfence_freelist_lock ->proc_subdir_lock ->&ent->pde_unload_lock ->proc_inum_ida.xa_lock ->sysfs_symlink_target_lock ->kernfs_idr_lock FD: 7 BD: 103 -.-.: &irq_desc_lock_class ->irq_controller_lock ->mask_lock ->&its->lock ->irq_resend_lock ->tmp_mask_lock FD: 21 BD: 14 +.+.: &desc->request_mutex ->&irq_desc_lock_class ->proc_subdir_lock ->&ent->pde_unload_lock ->proc_inum_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 ....: cpu_pm_notifier.lock FD: 14 BD: 27 +.+.: purge_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock FD: 1 BD: 3 +.+.: iort_msi_chip_lock FD: 2 BD: 2 ....: its_lock ->&its->lock FD: 1 BD: 2 ....: efi_mem_reserve_persistent_lock FD: 4 BD: 9 +.+.: lpi_range_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: syscore_ops_lock FD: 1 BD: 108 ....: &its->lock FD: 1 BD: 88 +.+.: cpuhp_state-down FD: 151 BD: 88 +.+.: cpuhp_state-up ->smpboot_threads_lock ->sparse_irq_lock ->&swhash->hlist_mutex ->pmus_lock ->&x->wait#4 ->&obj_hash[i].lock ->hrtimer_bases.lock ->wq_pool_mutex ->rcu_node_0 ->resource_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&rq->__lock ->fs_reclaim ->lock ->&root->kernfs_rwsem ->&x->wait#9 ->&k->list_lock ->bus_type_sem ->&k->k_lock ->&pcp->lock ->swap_slots_cache_mutex FD: 4 BD: 1 -.-.: timekeeper_lock ->tk_core.seq.seqcount FD: 3 BD: 3556 ----: tk_core.seq.seqcount ->&obj_hash[i].lock FD: 1 BD: 104 ....: irq_controller_lock FD: 7 BD: 88 ....: clockevents_lock ->tk_core.seq.seqcount ->tick_broadcast_lock ->jiffies_seq.seqcount FD: 3 BD: 89 -...: tick_broadcast_lock ->jiffies_lock FD: 1 BD: 91 -.-.: jiffies_seq.seqcount FD: 229 BD: 2 +.+.: clocksource_mutex ->cpu_hotplug_lock ->(console_sem).lock FD: 11 BD: 3570 -.-.: &base->lock ->&obj_hash[i].lock FD: 3 BD: 5 ....: batched_entropy_u64.lock ->crngs.lock FD: 121 BD: 89 +.+.: pmus_lock ->pcpu_alloc_mutex ->pool_lock#2 ->&obj_hash[i].lock ->&cpuctx_mutex ->fs_reclaim ->&k->list_lock ->lock ->&root->kernfs_rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->&c->lock ->&____s->seqcount ->&x->wait#9 ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#29 FD: 1 BD: 89 +.+.: &swhash->hlist_mutex FD: 1 BD: 90 +.+.: &cpuctx_mutex FD: 1 BD: 3 ....: tty_ldiscs_lock FD: 2 BD: 6 ....: kbd_event_lock ->led_lock FD: 1 BD: 7 ..-.: led_lock FD: 1 BD: 215 ....: console_owner_lock FD: 39 BD: 3 +.+.: init_task.alloc_lock ->init_fs.lock FD: 52 BD: 2 +.+.: acpi_ioremap_lock ->pool_lock#2 ->fs_reclaim ->&____s->seqcount ->free_vmap_area_lock ->vmap_area_lock FD: 1 BD: 15 ....: semaphore->lock FD: 1 BD: 14 +.+.: *(&acpi_gbl_reference_count_lock) FD: 13 BD: 3540 -.-.: hrtimer_bases.lock ->tk_core.seq.seqcount ->&obj_hash[i].lock FD: 1 BD: 761 ..-.: percpu_counters_lock FD: 33 BD: 2 +.+.: tomoyo_policy_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->&rq->__lock FD: 1419 BD: 4 ++++: pernet_ops_rwsem ->stack_depot_init_mutex ->crngs.lock ->net_rwsem ->proc_inum_ida.xa_lock ->pool_lock#2 ->proc_subdir_lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->sysctl_lock ->pcpu_alloc_mutex ->net_generic_ids.xa_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->&obj_hash[i].lock ->k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK ->nl_table_lock ->nl_table_wait.lock ->rtnl_mutex ->uevent_sock_mutex ->&net->rules_mod_lock ->slab_mutex ->batched_entropy_u32.lock ->percpu_counters_lock ->k-slock-AF_INET/1 ->&zone->lock ->cache_list_lock ->tk_core.seq.seqcount ->&k->list_lock ->lock ->&root->kernfs_rwsem ->running_helpers_waitq.lock ->&sn->pipefs_sb_lock ->krc.lock ->&rq->__lock ->pool_lock ->&s->s_inode_list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->nf_hook_mutex ->cpu_hotplug_lock ->hwsim_netgroup_ida.xa_lock ->nf_ct_ecache_mutex ->nf_log_mutex ->ipvs->est_mutex ->&base->lock ->__ip_vs_app_mutex ->&hashinfo->lock#2 ->&net->ipv6.ip6addrlbl_table.lock ->(console_sem).lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->wq_pool_mutex ->pcpu_lock ->&list->lock#4 ->&dir->lock#2 ->ptype_lock ->k-clock-AF_TIPC ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&this->receive_lock ->once_lock ->nf_ct_proto_mutex ->k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->conn_lock ->&call->waitq ->&rx->call_lock ->&rxnet->call_lock ->&n->list_lock ->rdma_nets.xa_lock ->devices_rwsem ->rtnl_mutex.wait_lock ->&p->pi_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&net->nsid_lock ->ebt_mutex ->nf_nat_proto_mutex ->&xt[i].mutex ->&nft_net->commit_mutex ->&lock->wait_lock ->&cfs_rq->removed.lock ->netns_bpf_mutex ->&x->wait#2 ->(&net->fs_probe_timer) ->&net->cells_lock ->(&net->cells_timer) ->(&net->fs_timer) ->bit_wait_table + i ->(wq_completion)kafsd ->&wq->mutex ->k-clock-AF_RXRPC ->&local->services_lock ->(wq_completion)krxrpcd ->rlock-AF_RXRPC ->&x->wait ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&ent->pde_unload_lock ->ovs_mutex ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&ovs_net->dp_notify_work) ->&srv->idr_lock ->rcu_node_0 ->&rnp->exp_wq[1] ->(work_completion)(&tn->work) ->&rnp->exp_wq[2] ->&tn->nametbl_lock ->&rnp->exp_wq[3] ->&rnp->exp_wq[0] ->(work_completion)(&ht->run_work) ->&ht->mutex ->(work_completion)(&(&c->work)->work) ->(wq_completion)krdsd ->(work_completion)(&rtn->rds_tcp_accept_w) ->rds_tcp_conn_lock ->loop_conns_lock ->(wq_completion)l2tp ->rcu_state.barrier_mutex ->(&rxnet->peer_keepalive_timer) ->(work_completion)(&rxnet->peer_keepalive_work) ->(&rxnet->service_conn_reap_timer) ->&x->wait#10 ->dev_base_lock ->lweventlist_lock ->napi_hash_lock ->netdev_unregistering_wq.lock ->&fn->fou_lock ->ipvs->sync_mutex ->hwsim_radio_lock ->pin_fs_lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key#3 ->&sb->s_type->i_lock_key#7 ->mount_lock ->(inetaddr_chain).rwsem ->inet6addr_chain.lock ->(work_completion)(&local->restart_work) ->&list->lock#16 ->&rdev->wiphy.mtx ->(work_completion)(&rfkill->uevent_work) ->(work_completion)(&rfkill->sync_work) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->&k->k_lock ->sysfs_symlink_target_lock ->subsys mutex#40 ->&x->wait#9 ->dpm_list_mtx ->&dev->power.lock ->deferred_probe_mutex ->device_links_lock ->&rfkill->lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->triggers_list_lock ->leds_list_lock ->(work_completion)(&rdev->wiphy_work) ->(work_completion)(&rdev->conn_work) ->(work_completion)(&rdev->event_work) ->(work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->(work_completion)(&(&rdev->background_cac_done_wk)->work) ->(work_completion)(&rdev->destroy_work) ->(work_completion)(&rdev->propagate_radar_detect_wk) ->(work_completion)(&rdev->propagate_cac_done_wk) ->(work_completion)(&rdev->mgmt_registrations_update_wk) ->(work_completion)(&rdev->background_cac_abort_wk) ->&sem->wait_lock ->uevent_sock_mutex.wait_lock ->subsys mutex#53 ->gdp_mutex ->(&local->sta_cleanup) ->rdma_nets_rwsem ->k-clock-AF_NETLINK ->&nlk->wait ->&wg->device_update_lock ->&bat_priv->forw_bcast_list_lock ->&bat_priv->forw_bat_list_lock ->&bat_priv->gw.list_lock ->(work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) ->&bat_priv->bat_v.ogm_buff_mutex ->&bat_priv->tvlv.container_list_lock ->&bat_priv->tvlv.handler_list_lock ->(work_completion)(&(&bat_priv->nc.work)->work) ->key#16 ->key#17 ->(work_completion)(&(&bat_priv->dat.work)->work) ->&hash->list_locks[i] ->(work_completion)(&(&bat_priv->bla.work)->work) ->key#19 ->(work_completion)(&(&bat_priv->mcast.work)->work) ->(work_completion)(&(&bat_priv->tt.work)->work) ->key#15 ->key#20 ->&bat_priv->tt.req_list_lock ->&bat_priv->tt.changes_list_lock ->&bat_priv->tt.roam_list_lock ->(work_completion)(&(&bat_priv->orig_work)->work) ->key#18 ->wq_mayday_lock ->&hn->hn_lock ->&pnettable->lock ->&pnetids_ndev->lock ->k-sk_lock-AF_INET6/1 ->&net->sctp.addr_wq_lock ->&rcu_state.expedited_wq ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->&sn->gssp_lock ->&cd->hash_lock ->(&net->can.stattimer) ->vmap_area_lock ->purge_vmap_area_lock ->stock_lock ->xfrm_state_gc_work ->&net->xfrm.xfrm_state_lock ->&sb->s_type->i_lock_key#23 ->rename_lock.seqcount ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->ip6_fl_lock ->(&net->ipv6.ip6_fib_timer) ->__ip_vs_mutex ->(&ipvs->dest_trash_timer) ->(work_completion)(&(&ipvs->expire_nodest_conn_work)->work) ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&ipvs->est_reload_work)->work) ->nfnl_subsys_ipset ->recent_lock ->hashlimit_mutex ->trans_gc_work ->&rnp->exp_lock ->rcu_state.exp_mutex ->nf_conntrack_mutex ->(work_completion)(&(&cnet->ecache.dwork)->work) ->tcp_metrics_lock ->k-clock-AF_INET ->(work_completion)(&net->xfrm.policy_hash_work) ->&net->xfrm.xfrm_policy_lock ->(work_completion)(&net->xfrm.state_hash_work) ->&xa->xa_lock#4 ->genl_sk_destructing_waitq.lock ->rcu_state.exp_mutex.wait_lock ->rcu_state.barrier_mutex.wait_lock ->quarantine_lock ->&meta->lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->wq_pool_mutex.wait_lock ->(wq_completion)phy40 ->(wq_completion)phy39 ->(wq_completion)bond0#19 ->(wq_completion)tipc_rcv#17 ->(wq_completion)tipc_send#17 ->(wq_completion)tipc_crypto#17 ->(wq_completion)phy32 ->(wq_completion)phy31 ->(wq_completion)bond0#15 ->(wq_completion)tipc_rcv#18 ->(wq_completion)tipc_send#18 ->(wq_completion)tipc_crypto#18 ->(wq_completion)phy44 ->(wq_completion)phy43 ->(wq_completion)bond0#21 ->(wq_completion)tipc_rcv#19 ->(wq_completion)tipc_send#19 ->(wq_completion)tipc_crypto#19 ->(wq_completion)phy46 ->(wq_completion)phy45 ->(wq_completion)bond0#22 ->(wq_completion)tipc_rcv#20 ->(wq_completion)tipc_send#20 ->(wq_completion)tipc_crypto#20 ->(wq_completion)phy48 ->(wq_completion)phy47 ->(wq_completion)bond0#23 ->pcpu_alloc_mutex.wait_lock ->(wq_completion)tipc_rcv#21 ->(wq_completion)tipc_send#21 ->(wq_completion)tipc_crypto#21 ->(wq_completion)phy50 ->(wq_completion)phy49 ->(wq_completion)bond0#24 ->(wq_completion)tipc_rcv#22 ->(wq_completion)tipc_send#22 ->(wq_completion)tipc_crypto#22 ->(wq_completion)phy34 ->(wq_completion)phy33 ->(wq_completion)bond0#16 ->(wq_completion)tipc_rcv#23 ->(wq_completion)tipc_send#23 ->(wq_completion)tipc_crypto#23 ->(wq_completion)phy54 ->(wq_completion)phy53 ->(wq_completion)bond0#27 ->(wq_completion)tipc_rcv#24 ->(wq_completion)tipc_send#24 ->(wq_completion)tipc_crypto#24 ->key ->(wq_completion)tipc_rcv#25 ->(wq_completion)tipc_send#25 ->(wq_completion)tipc_rcv#26 ->(wq_completion)tipc_send#26 ->(wq_completion)tipc_rcv#27 ->(wq_completion)tipc_send#27 ->(wq_completion)tipc_rcv#28 ->(wq_completion)tipc_send#28 ->(wq_completion)tipc_rcv#29 ->(wq_completion)tipc_send#29 ->(wq_completion)tipc_rcv#30 ->(wq_completion)tipc_send#30 ->(wq_completion)tipc_rcv#31 ->(wq_completion)tipc_send#31 ->(wq_completion)tipc_rcv#32 ->(wq_completion)tipc_send#32 ->(wq_completion)tipc_crypto#25 ->key#24 ->(wq_completion)tipc_crypto#26 ->(wq_completion)tipc_crypto#27 ->(wq_completion)tipc_crypto#28 ->(wq_completion)tipc_crypto#29 ->(wq_completion)tipc_crypto#30 ->(wq_completion)tipc_crypto#31 ->(wq_completion)tipc_crypto#32 ->nf_ct_proto_mutex.wait_lock ->ovs_mutex.wait_lock ->(wq_completion)tipc_rcv#33 ->(wq_completion)tipc_send#33 ->(wq_completion)tipc_rcv#34 ->(wq_completion)tipc_send#34 ->(wq_completion)tipc_rcv#35 ->(wq_completion)tipc_send#35 ->(wq_completion)tipc_rcv#36 ->(wq_completion)tipc_send#36 ->(wq_completion)tipc_rcv#37 ->(wq_completion)tipc_send#37 ->(wq_completion)tipc_rcv#38 ->(wq_completion)tipc_send#38 ->(wq_completion)tipc_rcv#39 ->(wq_completion)tipc_send#39 ->(wq_completion)tipc_rcv#40 ->(wq_completion)tipc_send#40 ->(wq_completion)tipc_rcv#41 ->(wq_completion)tipc_send#41 ->(wq_completion)tipc_rcv#42 ->(wq_completion)tipc_send#42 ->(wq_completion)tipc_rcv#43 ->(wq_completion)tipc_send#43 ->(wq_completion)tipc_rcv#44 ->(wq_completion)tipc_send#44 ->(wq_completion)tipc_rcv#45 ->(wq_completion)tipc_send#45 ->(wq_completion)tipc_rcv#46 ->(wq_completion)tipc_send#46 ->(wq_completion)tipc_rcv#47 ->(wq_completion)tipc_send#47 ->(wq_completion)tipc_rcv#48 ->(wq_completion)tipc_send#48 ->(wq_completion)tipc_rcv#49 ->(wq_completion)tipc_send#49 ->(wq_completion)tipc_rcv#50 ->(wq_completion)tipc_send#50 ->(wq_completion)tipc_rcv#51 ->(wq_completion)tipc_send#51 ->(wq_completion)tipc_rcv#52 ->(wq_completion)tipc_send#52 ->(wq_completion)tipc_rcv#53 ->(wq_completion)tipc_send#53 ->(wq_completion)tipc_rcv#54 ->(wq_completion)tipc_send#54 ->(wq_completion)tipc_rcv#55 ->(wq_completion)tipc_send#55 ->(wq_completion)tipc_rcv#56 ->(wq_completion)tipc_send#56 ->(wq_completion)tipc_rcv#57 ->(wq_completion)tipc_send#57 ->(wq_completion)tipc_rcv#58 ->(wq_completion)tipc_send#58 ->(wq_completion)tipc_crypto#33 ->(wq_completion)tipc_crypto#34 ->(wq_completion)tipc_crypto#35 ->(wq_completion)tipc_crypto#36 ->(wq_completion)tipc_crypto#37 ->(wq_completion)tipc_crypto#38 ->(wq_completion)tipc_crypto#39 ->(wq_completion)tipc_crypto#40 ->(wq_completion)tipc_crypto#41 ->(wq_completion)tipc_crypto#42 ->(wq_completion)tipc_crypto#43 ->(wq_completion)tipc_crypto#44 ->(wq_completion)tipc_crypto#45 ->(wq_completion)tipc_crypto#46 ->(wq_completion)tipc_crypto#47 ->(wq_completion)tipc_crypto#48 ->(wq_completion)tipc_crypto#49 ->(wq_completion)tipc_crypto#50 ->(wq_completion)tipc_crypto#51 ->(wq_completion)tipc_crypto#52 ->(wq_completion)tipc_crypto#53 ->(wq_completion)tipc_crypto#54 ->(wq_completion)tipc_crypto#55 ->(wq_completion)tipc_crypto#56 ->(wq_completion)tipc_crypto#57 ->(wq_completion)tipc_crypto#58 ->(wq_completion)phy52 ->(wq_completion)phy51 ->(wq_completion)bond0#25 ->nf_hook_mutex.wait_lock ->(wq_completion)tipc_rcv#59 ->(wq_completion)tipc_send#59 ->(wq_completion)tipc_rcv#60 ->(wq_completion)tipc_send#60 ->(wq_completion)tipc_rcv#61 ->(wq_completion)tipc_send#61 ->(wq_completion)tipc_rcv#62 ->(wq_completion)tipc_send#62 ->(wq_completion)tipc_rcv#63 ->(wq_completion)tipc_send#63 ->(wq_completion)tipc_rcv#64 ->(wq_completion)tipc_send#64 ->(wq_completion)tipc_rcv#65 ->(wq_completion)tipc_send#65 ->(wq_completion)tipc_rcv#66 ->(wq_completion)tipc_send#66 ->(wq_completion)tipc_rcv#67 ->(wq_completion)tipc_send#67 ->(wq_completion)tipc_rcv#68 ->(wq_completion)tipc_send#68 ->(wq_completion)tipc_rcv#69 ->(wq_completion)tipc_send#69 ->(wq_completion)tipc_rcv#70 ->(wq_completion)tipc_send#70 ->(wq_completion)tipc_rcv#71 ->(wq_completion)tipc_send#71 ->(wq_completion)tipc_rcv#72 ->(wq_completion)tipc_send#72 ->(wq_completion)tipc_rcv#73 ->(wq_completion)tipc_send#73 ->(wq_completion)tipc_rcv#74 ->(wq_completion)tipc_send#74 ->(wq_completion)tipc_rcv#75 ->(wq_completion)tipc_send#75 ->(wq_completion)tipc_rcv#76 ->(wq_completion)tipc_send#76 ->(wq_completion)tipc_rcv#77 ->(wq_completion)tipc_send#77 ->(wq_completion)tipc_rcv#78 ->(wq_completion)tipc_send#78 ->(wq_completion)tipc_rcv#79 ->(wq_completion)tipc_send#79 ->(wq_completion)tipc_rcv#80 ->(wq_completion)tipc_send#80 ->(wq_completion)tipc_rcv#81 ->(wq_completion)tipc_send#81 ->(wq_completion)tipc_rcv#82 ->(wq_completion)tipc_send#82 ->(wq_completion)tipc_rcv#83 ->(wq_completion)tipc_send#83 ->(wq_completion)tipc_crypto#59 ->(wq_completion)tipc_crypto#60 ->(wq_completion)tipc_crypto#61 ->(wq_completion)tipc_crypto#62 ->(wq_completion)tipc_crypto#63 ->(wq_completion)tipc_crypto#64 ->(wq_completion)tipc_crypto#65 ->(wq_completion)tipc_crypto#66 ->(wq_completion)tipc_crypto#67 ->(wq_completion)tipc_crypto#68 ->(wq_completion)tipc_crypto#69 ->(wq_completion)tipc_crypto#70 ->(wq_completion)tipc_crypto#71 ->(wq_completion)tipc_crypto#72 ->(wq_completion)tipc_crypto#73 ->(wq_completion)tipc_crypto#74 ->(wq_completion)tipc_crypto#75 ->(wq_completion)tipc_crypto#76 ->(wq_completion)tipc_crypto#77 ->(wq_completion)tipc_crypto#78 ->(wq_completion)tipc_crypto#79 ->(wq_completion)tipc_crypto#80 ->(wq_completion)tipc_crypto#81 ->(wq_completion)tipc_crypto#82 ->(wq_completion)tipc_crypto#83 ->(wq_completion)phy42 ->kn->active#4 ->(wq_completion)phy41 ->(wq_completion)bond0#20 ->(wq_completion)tipc_rcv#84 ->(wq_completion)tipc_send#84 ->(wq_completion)tipc_rcv#85 ->(wq_completion)tipc_send#85 ->(wq_completion)tipc_rcv#86 ->(wq_completion)tipc_send#86 ->(wq_completion)tipc_rcv#87 ->(wq_completion)tipc_send#87 ->(wq_completion)tipc_rcv#88 ->(wq_completion)tipc_send#88 ->(wq_completion)tipc_rcv#89 ->(wq_completion)tipc_send#89 ->(wq_completion)tipc_rcv#90 ->(wq_completion)tipc_send#90 ->(wq_completion)tipc_rcv#91 ->(wq_completion)tipc_send#91 ->(wq_completion)tipc_rcv#92 ->(wq_completion)tipc_send#92 ->(wq_completion)tipc_rcv#93 ->(wq_completion)tipc_send#93 ->(wq_completion)tipc_rcv#94 ->(wq_completion)tipc_send#94 ->(wq_completion)tipc_rcv#95 ->(wq_completion)tipc_send#95 ->(wq_completion)tipc_rcv#96 ->(wq_completion)tipc_send#96 ->(wq_completion)tipc_rcv#97 ->(wq_completion)tipc_send#97 ->(wq_completion)tipc_rcv#98 ->(wq_completion)tipc_send#98 ->(wq_completion)tipc_rcv#99 ->(wq_completion)tipc_send#99 ->(wq_completion)tipc_rcv#100 ->(wq_completion)tipc_send#100 ->(wq_completion)tipc_rcv#101 ->(wq_completion)tipc_send#101 ->(wq_completion)tipc_rcv#102 ->(wq_completion)tipc_send#102 ->(wq_completion)tipc_rcv#103 ->(wq_completion)tipc_send#103 ->(wq_completion)tipc_rcv#104 ->(wq_completion)tipc_send#104 ->(wq_completion)tipc_rcv#105 ->(wq_completion)tipc_send#105 ->(wq_completion)tipc_rcv#106 ->(wq_completion)tipc_send#106 ->(wq_completion)tipc_rcv#107 ->(wq_completion)tipc_send#107 ->(wq_completion)tipc_rcv#108 ->(wq_completion)tipc_send#108 ->(wq_completion)tipc_rcv#109 ->(wq_completion)tipc_send#109 ->(wq_completion)tipc_rcv#110 ->(wq_completion)tipc_send#110 ->(wq_completion)tipc_rcv#111 ->(wq_completion)tipc_send#111 ->(wq_completion)tipc_rcv#112 ->(wq_completion)tipc_send#112 ->(wq_completion)tipc_crypto#84 ->(wq_completion)tipc_crypto#85 ->(wq_completion)tipc_crypto#86 ->(wq_completion)tipc_crypto#87 ->(wq_completion)tipc_crypto#88 ->(wq_completion)tipc_crypto#89 ->(wq_completion)tipc_crypto#90 ->(wq_completion)tipc_crypto#91 ->(wq_completion)tipc_crypto#92 ->(wq_completion)tipc_crypto#93 ->(wq_completion)tipc_crypto#94 ->(wq_completion)tipc_crypto#95 ->(wq_completion)tipc_crypto#96 ->(wq_completion)tipc_crypto#97 ->(wq_completion)tipc_crypto#98 ->(wq_completion)tipc_crypto#99 ->(wq_completion)tipc_crypto#100 ->(wq_completion)tipc_crypto#101 ->(wq_completion)tipc_crypto#102 ->(wq_completion)tipc_crypto#103 ->(wq_completion)tipc_crypto#104 ->(wq_completion)tipc_crypto#105 ->(wq_completion)tipc_crypto#106 ->(wq_completion)tipc_crypto#107 ->(wq_completion)tipc_crypto#108 ->(wq_completion)tipc_crypto#109 ->(wq_completion)tipc_crypto#110 ->(wq_completion)tipc_crypto#111 ->(wq_completion)tipc_crypto#112 ->(wq_completion)tipc_rcv#113 ->(wq_completion)tipc_send#113 ->(wq_completion)tipc_rcv#114 ->(wq_completion)tipc_send#114 ->(wq_completion)tipc_rcv#115 ->(wq_completion)tipc_send#115 ->(wq_completion)tipc_rcv#116 ->(wq_completion)tipc_send#116 ->(wq_completion)tipc_rcv#117 ->(wq_completion)tipc_send#117 ->(wq_completion)tipc_rcv#118 ->(wq_completion)tipc_send#118 ->(wq_completion)tipc_rcv#119 ->(wq_completion)tipc_send#119 ->(wq_completion)tipc_rcv#120 ->(wq_completion)tipc_send#120 ->(wq_completion)tipc_rcv#121 ->(wq_completion)tipc_send#121 ->(wq_completion)tipc_rcv#122 ->(wq_completion)tipc_send#122 ->(wq_completion)tipc_rcv#123 ->(wq_completion)tipc_send#123 ->(wq_completion)tipc_rcv#124 ->(wq_completion)tipc_send#124 ->(wq_completion)tipc_rcv#125 ->(wq_completion)tipc_send#125 ->(wq_completion)tipc_rcv#126 ->(wq_completion)tipc_send#126 ->(wq_completion)tipc_rcv#127 ->(wq_completion)tipc_send#127 ->(wq_completion)tipc_rcv#128 ->(wq_completion)tipc_send#128 ->(wq_completion)tipc_rcv#129 ->(wq_completion)tipc_send#129 ->(wq_completion)tipc_rcv#130 ->(wq_completion)tipc_send#130 ->(wq_completion)tipc_rcv#131 ->(wq_completion)tipc_send#131 ->(wq_completion)tipc_rcv#132 ->(wq_completion)tipc_send#132 ->(wq_completion)tipc_rcv#133 ->(wq_completion)tipc_send#133 ->(wq_completion)tipc_rcv#134 ->(wq_completion)tipc_send#134 ->(wq_completion)tipc_rcv#135 ->(wq_completion)tipc_send#135 ->(wq_completion)tipc_rcv#136 ->(wq_completion)tipc_send#136 ->(wq_completion)tipc_rcv#137 ->(wq_completion)tipc_send#137 ->(wq_completion)tipc_rcv#138 ->(wq_completion)tipc_send#138 ->(wq_completion)tipc_crypto#113 ->(wq_completion)tipc_crypto#114 ->(wq_completion)tipc_crypto#115 ->(wq_completion)tipc_crypto#116 ->(wq_completion)tipc_crypto#117 ->(wq_completion)tipc_crypto#118 ->(wq_completion)tipc_crypto#119 ->(wq_completion)tipc_crypto#120 ->(wq_completion)tipc_crypto#121 ->(wq_completion)tipc_crypto#122 ->(wq_completion)tipc_crypto#123 ->(wq_completion)tipc_crypto#124 ->(wq_completion)tipc_crypto#125 ->(wq_completion)tipc_crypto#126 ->(wq_completion)tipc_crypto#127 ->(wq_completion)tipc_crypto#128 ->(wq_completion)tipc_crypto#129 ->(wq_completion)tipc_crypto#130 ->(wq_completion)tipc_crypto#131 ->(wq_completion)tipc_crypto#132 ->(wq_completion)tipc_crypto#133 ->(wq_completion)tipc_crypto#134 ->(wq_completion)tipc_crypto#135 ->(wq_completion)tipc_crypto#136 ->(wq_completion)tipc_crypto#137 ->(wq_completion)tipc_crypto#138 ->(wq_completion)tipc_rcv#139 ->(wq_completion)tipc_send#139 ->(wq_completion)tipc_rcv#140 ->(wq_completion)tipc_send#140 ->(wq_completion)tipc_crypto#139 ->(wq_completion)tipc_crypto#140 ->(wq_completion)phy56 ->(wq_completion)phy55 ->(wq_completion)bond0#26 ->(wq_completion)tipc_rcv#141 ->(wq_completion)tipc_send#141 ->(wq_completion)tipc_crypto#141 ->(wq_completion)phy62 ->(wq_completion)phy61 ->(wq_completion)bond0#30 ->(wq_completion)tipc_rcv#142 ->(wq_completion)tipc_send#142 ->(wq_completion)tipc_rcv#143 ->(wq_completion)tipc_send#143 ->(wq_completion)tipc_rcv#144 ->(wq_completion)tipc_send#144 ->(wq_completion)tipc_crypto#142 ->(wq_completion)tipc_crypto#143 ->(wq_completion)tipc_crypto#144 ->(wq_completion)phy60 ->(wq_completion)phy59 ->(wq_completion)phy38 ->(wq_completion)phy37 ->(wq_completion)phy64 ->(wq_completion)phy63 ->(wq_completion)bond0#29 ->(wq_completion)bond0#18 ->(wq_completion)bond0#31 ->(wq_completion)tipc_rcv#145 ->(wq_completion)tipc_send#145 ->(wq_completion)tipc_rcv#146 ->(wq_completion)tipc_send#146 ->(wq_completion)tipc_crypto#145 ->(wq_completion)tipc_crypto#146 ->(wq_completion)phy66 ->(wq_completion)phy65 ->(wq_completion)phy58 ->(wq_completion)phy57 ->(wq_completion)bond0#32 ->(wq_completion)bond0#28 ->(wq_completion)tipc_rcv#147 ->(wq_completion)tipc_send#147 ->(wq_completion)tipc_crypto#147 ->(wq_completion)phy76 ->(wq_completion)phy75 ->(wq_completion)bond0#37 FD: 25 BD: 49 +.+.: stack_depot_init_mutex ->&rq->__lock ->stack_depot_init_mutex.wait_lock ->&cfs_rq->removed.lock FD: 35 BD: 82 ++++: net_rwsem ->&list->lock#2 ->&rq->__lock ->pool_lock#2 ->&obj_hash[i].lock ->nl_table_lock ->&____s->seqcount ->nl_table_wait.lock ->quarantine_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->&c->lock ->net_rwsem.wait_lock ->&n->list_lock FD: 11 BD: 111 ..-.: proc_inum_ida.xa_lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&____s->seqcount FD: 849 BD: 44 +.+.: rtnl_mutex ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->fs_reclaim ->pcpu_alloc_mutex ->&xa->xa_lock#4 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#20 ->&dir->lock#2 ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->nl_table_lock ->nl_table_wait.lock ->net_rwsem ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->krc.lock ->&rq->__lock ->stack_depot_init_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->wq_pool_mutex ->crngs.lock ->pool_lock ->lweventlist_lock ->rtnl_mutex.wait_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->quarantine_lock ->&k->k_lock ->param_lock ->(console_sem).lock ->&rdev->wiphy.mtx ->&base->lock ->subsys mutex#55 ->&sdata->sec_mtx ->&cfs_rq->removed.lock ->&local->iflist_mtx#2 ->lock#7 ->failover_lock ->&tn->lock ->&idev->mc_lock ->&ndev->lock ->&pnettable->lock ->smc_ib_devices.mutex ->&(&net->nexthop.notifier_chain)->rwsem ->reg_requests_lock ->reg_pending_beacons_lock ->devnet_rename_sem ->&x->wait#2 ->&nft_net->commit_mutex ->&ent->pde_unload_lock ->target_list_lock ->rlock-AF_NETLINK ->(inetaddr_validator_chain).rwsem ->(inetaddr_chain).rwsem ->_xmit_LOOPBACK ->netpoll_srcu ->&in_dev->mc_tomb_lock ->&im->lock ->fib_info_lock ->cbs_list_lock ->(inet6addr_validator_chain).rwsem ->&net->ipv6.addrconf_hash_lock ->&ifa->lock ->&tb->tb6_lock ->&n->list_lock ->&dev_addr_list_lock_key ->napi_hash_lock ->lapb_list_lock ->remove_cache_srcu ->x25_neigh_list_lock ->console_owner_lock ->console_owner ->_xmit_ETHER ->_xmit_SLIP ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->&cma->lock ->cma_mutex ->rcu_node_0 ->&sem->wait_lock ->&rfkill->lock ->_xmit_VOID ->_xmit_X25 ->&lapbeth->up_lock ->&lapb->lock ->class ->(&tbl->proxy_timer) ->&dev->tx_global_lock ->&rnp->exp_wq[2] ->&sch->q.lock ->&rnp->exp_wq[3] ->&dir->lock ->&ul->lock#2 ->&n->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&wpan_dev->association_lock ->dev_addr_sem ->_xmit_IEEE802154 ->&nr_netdev_addr_lock_key ->listen_lock ->&r->consumer_lock ->&mm->mmap_lock ->pcpu_lock ->(switchdev_blocking_notif_chain).rwsem ->&br->hash_lock ->nf_hook_mutex ->j1939_netdev_lock ->key ->percpu_counters_lock ->&bat_priv->tvlv.handler_list_lock ->&bat_priv->tvlv.container_list_lock ->&bat_priv->softif_vlan_list_lock ->key#15 ->&bat_priv->tt.changes_list_lock ->kernfs_idr_lock ->&rnp->exp_wq[0] ->noop_qdisc.q.lock ->&rnp->exp_wq[1] ->tk_core.seq.seqcount ->&wq->mutex ->hrtimer_bases.lock ->init_lock ->deferred_lock ->&br->lock ->&pn->hash_lock ->&hard_iface->bat_iv.ogm_buff_mutex ->ptype_lock ->_xmit_NONE ->lock#9 ->&hsr->list_lock ->&meta->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&xa->xa_lock#18 ->&dev_addr_list_lock_key#3/1 ->req_lock ->&x->wait#11 ->subsys mutex#75 ->bpf_devs_lock ->(work_completion)(&(&devlink_port->type_warn_dw)->work) ->&devlink_port->type_lock ->&vn->sock_lock ->mount_lock ->&wg->device_update_lock ->_xmit_SIT ->&bridge_netdev_addr_lock_key/1 ->_xmit_TUNNEL ->_xmit_IPGRE ->_xmit_TUNNEL6 ->&dev_addr_list_lock_key/1 ->&dev_addr_list_lock_key#2/1 ->_xmit_ETHER/1 ->&nn->netlink_tap_lock ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&ipvlan->addrs_lock ->&macsec_netdev_addr_lock_key/1 ->key#19 ->&bat_priv->tt.commit_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->k-slock-AF_INET/1 ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&____s->seqcount#2 ->&ul->lock ->&tun->lock ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->__ip_vs_mutex ->flowtable_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&hwstats->hwsdev_list_lock ->&net->xdp.lock ->mirred_list_lock ->&idev->mc_report_lock ->&sb->s_type->i_lock_key#23 ->&dentry->d_lock ->rename_lock.seqcount ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&pnn->pndevs.lock ->&pnn->routes.lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->uevent_sock_mutex.wait_lock ->pcpu_alloc_mutex.wait_lock ->&net->xfrm.xfrm_state_lock ->&net->xfrm.xfrm_policy_lock ->&sb->s_type->i_lock_key#7 ->(work_completion)(&wdev->disconnect_wk) ->(work_completion)(&wdev->pmsr_free_wk) ->&rdev->dev_wait ->&fq->lock ->&app->lock#2 ->(&app->join_timer)#2 ->(&app->periodic_timer) ->&list->lock#11 ->(&app->join_timer) ->&app->lock ->&list->lock#10 ->(work_completion)(&(&priv->scan_result)->work) ->(work_completion)(&(&priv->connect)->work) ->(&hsr->prune_timer) ->(&hsr->announce_timer) ->key#18 ->&bat_priv->forw_bcast_list_lock ->&bat_priv->forw_bat_list_lock ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->(&pmctx->ip6_mc_router_timer) ->(&pmctx->ip4_mc_router_timer) ->(work_completion)(&ht->run_work) ->&ht->mutex ->&br->multicast_lock ->(work_completion)(&(&br->gc_work)->work) ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&table->hash[i].lock ->k-clock-AF_INET6 ->k-clock-AF_INET ->&r->consumer_lock#2 ->&wg->socket_update_lock ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->arp_work)->work) ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->(work_completion)(&(&bond->slave_arr_work)->work) ->(&br->hello_timer) ->(&br->topology_change_timer) ->(&br->tcn_timer) ->(&brmctx->ip4_mc_router_timer) ->(&brmctx->ip4_other_query.timer) ->(&brmctx->ip4_other_query.delay_timer) ->(&brmctx->ip4_own_query.timer) ->(&brmctx->ip6_mc_router_timer) ->(&brmctx->ip6_other_query.timer) ->(&brmctx->ip6_other_query.delay_timer) ->(&brmctx->ip6_own_query.timer) ->raw_notifier_lock ->bcm_notifier_lock ->isotp_notifier_lock ->(work_completion)(&port->bc_work) ->(work_completion)(&port->wq) ->(work_completion)(&(&slave->notify_work)->work) ->_xmit_NETROM#2 ->&pmc->lock ->(&mp->timer) ->(work_completion)(&br->mcast_gc_work) ->rcu_state.barrier_mutex ->&rcu_state.expedited_wq ->&caifn->caifdevs.lock ->&net->rules_mod_lock ->(&mrt->ipmr_expire_timer) ->rcu_state.exp_mutex.wait_lock ->&rnp->exp_lock ->stock_lock ->rcu_state.exp_mutex ->reg_indoor_lock ->netlbl_unlhsh_lock ->nr_list_lock ->nr_neigh_list_lock ->&bpq_netdev_addr_lock_key ->gdp_mutex.wait_lock ->&lock->wait_lock ->key#22 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->dev_pm_qos_sysfs_mtx.wait_lock ->net_rwsem.wait_lock ->team->team_lock_key#33 ->team->team_lock_key#34 ->team->team_lock_key#35 ->team->team_lock_key#36 ->team->team_lock_key#38 FD: 39 BD: 216 +.+.: lock ->kernfs_idr_lock ->cgroup_idr_lock ->pidmap_lock ->drm_minor_lock ->&file_private->table_lock ->&q->queue_lock ->&group->inotify_data.idr_lock ->map_idr_lock ->prog_idr_lock FD: 13 BD: 227 +.+.: kernfs_idr_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 84 BD: 220 ++++: &root->kernfs_rwsem ->&root->kernfs_iattr_rwsem ->kernfs_idr_lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&rq->__lock ->&cfs_rq->removed.lock ->quarantine_lock ->rcu_node_0 ->inode_hash_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->kernfs_rename_lock ->&n->list_lock ->&sem->wait_lock ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 ->&meta->lock ->&xa->xa_lock#5 ->stock_lock ->&____s->seqcount#2 ->&p->pi_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&base->lock FD: 1 BD: 4 ++++: file_systems_lock FD: 50 BD: 224 ++++: &root->kernfs_iattr_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->iattr_mutex ->&sem->wait_lock ->tk_core.seq.seqcount ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: dq_list_lock FD: 5 BD: 45 +.+.: sb_lock ->unnamed_dev_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 85 BD: 1 +.+.: &type->s_umount_key/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start FD: 1 BD: 39 +.+.: list_lrus_mutex FD: 1 BD: 46 ....: unnamed_dev_ida.xa_lock FD: 1 BD: 18 +.+.: &sbinfo->stat_lock FD: 46 BD: 125 +.+.: &sb->s_type->i_lock_key ->&dentry->d_lock ->&xa->xa_lock#9 FD: 1 BD: 328 +.+.: &s->s_inode_list_lock FD: 36 BD: 411 +.+.: &dentry->d_lock ->&wq ->&dentry->d_lock/1 ->&obj_hash[i].lock ->pool_lock#2 ->&wq#2 ->&lru->node[i].lock ->sysctl_lock ->&dentry->d_lock/2 ->&p->pi_lock FD: 2 BD: 29 ....: mnt_id_ida.xa_lock ->pool_lock#2 FD: 40 BD: 121 +.+.: mount_lock ->mount_lock.seqcount ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 38 BD: 121 +.+.: mount_lock.seqcount ->&new_ns->poll ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 75 BD: 1 +.+.: &type->s_umount_key#2/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 112 +.+.: &sb->s_type->i_lock_key#2 ->&dentry->d_lock FD: 1 BD: 2 ..-.: ucounts_lock FD: 38 BD: 130 +.+.: init_fs.lock ->init_fs.seq.seqcount ->&dentry->d_lock FD: 1 BD: 124 +.+.: init_fs.seq.seqcount FD: 2 BD: 90 -.-.: jiffies_lock ->jiffies_seq.seqcount FD: 25 BD: 1 -.-.: log_wait.lock ->&p->pi_lock FD: 83 BD: 1 +.+.: &type->s_umount_key#3/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 46 BD: 287 +.+.: &sb->s_type->i_lock_key#3 ->&dentry->d_lock ->&xa->xa_lock#9 FD: 1 BD: 112 ++++: proc_subdir_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#4/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&sb->s_type->i_lock_key#4 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#4 ->&dentry->d_lock FD: 31 BD: 92 ..-.: cgroup_file_kn_lock ->kernfs_notify_lock ->&obj_hash[i].lock ->&base->lock FD: 33 BD: 90 ..-.: css_set_lock ->cgroup_file_kn_lock ->&p->pi_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 2 BD: 217 +...: cgroup_idr_lock ->pool_lock#2 FD: 52 BD: 89 +.+.: cpuset_mutex ->callback_lock ->jump_label_mutex ->&p->pi_lock ->&p->alloc_lock ->cpuset_attach_wq.lock FD: 1 BD: 90 ....: callback_lock FD: 58 BD: 16 +.+.: blkcg_pol_mutex ->pcpu_alloc_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 16 +.+.: devcgroup_mutex FD: 30 BD: 89 +.+.: freezer_mutex ->freezer_lock ->rcu_node_0 ->&rq->__lock ->freezer_mutex.wait_lock ->&cfs_rq->removed.lock FD: 44 BD: 51 +.+.: rcu_state.exp_mutex ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&obj_hash[i].lock ->&rnp->exp_wq[2] ->&rq->__lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[3] ->pool_lock#2 ->&cfs_rq->removed.lock ->rcu_state.exp_mutex.wait_lock ->key ->pcpu_lock ->percpu_counters_lock ->stock_lock FD: 38 BD: 93 +.+.: rcu_state.exp_wake_mutex ->rcu_node_0 ->&rnp->exp_lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&rnp->exp_wq[3] ->&rq->__lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.exp_wake_mutex.wait_lock FD: 1 BD: 94 +.+.: &rnp->exp_lock FD: 25 BD: 95 ....: &rnp->exp_wq[0] ->&p->pi_lock FD: 25 BD: 96 ....: &rnp->exp_wq[1] ->&p->pi_lock FD: 1 BD: 93 ....: init_sighand.siglock FD: 1 BD: 3 +.+.: init_files.file_lock FD: 26 BD: 227 ....: pidmap_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&p->pi_lock FD: 130 BD: 88 ++++: cgroup_threadgroup_rwsem ->css_set_lock ->&p->pi_lock ->tk_core.seq.seqcount ->tasklist_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->&sighand->siglock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&x->wait#2 ->inode_hash_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#30 ->&root->kernfs_iattr_rwsem ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&c->lock ->cpuset_mutex ->freezer_mutex ->&p->alloc_lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->cgroup_threadgroup_rwsem.waiters.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->freezer_mutex.wait_lock ->stock_lock ->remove_cache_srcu ->(console_sem).lock ->kernfs_pr_cont_lock ->cgroup_file_kn_lock FD: 24 BD: 1113 -.-.: &p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 61 BD: 92 .+.+: tasklist_lock ->init_sighand.siglock ->&sighand->siglock ->&pid->wait_pidfd ->&obj_hash[i].lock ->quarantine_lock ->stock_lock ->&p->alloc_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&base->lock FD: 1 BD: 3515 -.-.: &per_cpu_ptr(group->pcpu, cpu)->seq FD: 1 BD: 1 ....: (kthreadd_done).wait.lock FD: 37 BD: 97 ....: &sighand->siglock ->&sig->wait_chldexit ->input_pool.lock ->&(&sig->stats_lock)->lock ->&p->pi_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->hrtimer_bases.lock ->&obj_hash[i].lock ->&sighand->signalfd_wqh ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&tty->ctrl.lock ->&rq->__lock ->stock_lock ->&n->list_lock FD: 45 BD: 128 +.+.: &p->alloc_lock ->&____s->seqcount#2 ->init_fs.lock ->&fs->lock ->&x->wait ->cpu_asid_lock ->&x->wait#25 ->&newf->file_lock FD: 1 BD: 3560 .-.-: &____s->seqcount#2 FD: 48 BD: 705 +.+.: fs_reclaim ->mmu_notifier_invalidate_range_start ->&mapping->i_mmap_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->icc_bw_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock FD: 32 BD: 728 +.+.: mmu_notifier_invalidate_range_start ->dma_fence_map ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 104 +.+.: kthread_create_lock FD: 25 BD: 147 ....: &x->wait ->&p->pi_lock FD: 32 BD: 1 +.+.: sched_map-wait-type-override ->&pool->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&__ctx->lock ->rcu_node_0 FD: 25 BD: 655 ....: (worker)->lock ->&p->pi_lock FD: 34 BD: 92 +.+.: wq_pool_attach_mutex ->&p->pi_lock ->&x->wait#7 ->&pool->lock ->&rq->__lock ->wq_pool_attach_mutex.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 655 ..-.: wq_mayday_lock FD: 1 BD: 91 ....: &xa->xa_lock FD: 30 BD: 1 +.-.: (&pool->mayday_timer) ->&pool->lock ->&obj_hash[i].lock ->&base->lock FD: 52 BD: 1 +.+.: (wq_completion)rcu_gp ->(work_completion)(&rnp->exp_poll_wq) ->(work_completion)(&(&ssp->srcu_sup->work)->work) ->(work_completion)(&sdp->work) ->(work_completion)(&rew->rew_work) ->&rq->__lock FD: 31 BD: 2 +.+.: (work_completion)(&rnp->exp_poll_wq) ->&rnp->exp_poll_lock FD: 4 BD: 1 -.-.: (null) ->tk_core.seq.seqcount FD: 12 BD: 1 +.-.: (&wq_watchdog_timer) ->&obj_hash[i].lock ->&base->lock FD: 534 BD: 1 +.+.: (wq_completion)events_unbound ->(work_completion)(&(&kfence_timer)->work) ->(work_completion)(&entry->work) ->(next_reseed).work ->(stats_flush_dwork).work ->(work_completion)(&sub_info->work) ->deferred_probe_work ->(work_completion)(&barr->work) ->connector_reaper_work ->(reaper_work).work ->(work_completion)(&rdev->wiphy_work) ->(work_completion)(&port->bc_work) ->(work_completion)(&map->work) ->&rq->__lock ->(work_completion)(&pool->idle_cull_work) FD: 230 BD: 2 +.+.: (work_completion)(&(&kfence_timer)->work) ->cpu_hotplug_lock ->allocation_wait.lock ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 25 BD: 3 -.-.: allocation_wait.lock ->&p->pi_lock FD: 3 BD: 3635 ..-.: batched_entropy_u8.lock ->crngs.lock FD: 1 BD: 3620 ..-.: kfence_freelist_lock FD: 1 BD: 694 ..-.: &meta->lock FD: 48 BD: 1 +.+.: rcu_tasks.tasks_gp_mutex ->&obj_hash[i].lock ->&base->lock ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(sdp, lock) ->tasks_rcu_exit_srcu ->&rq->__lock ->&x->wait#3 ->kernel/rcu/tasks.h:152 ->rcu_tasks__percpu.cbs_pcpu_lock ->&x->wait#2 ->(&timer.timer) ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks.cbs_gbl_lock FD: 229 BD: 1 +.+.: rcu_tasks_trace.tasks_gp_mutex ->cpu_hotplug_lock ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&x->wait#2 ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->(&timer.timer) ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks_trace.cbs_gbl_lock FD: 12 BD: 3 ..-.: rcu_tasks__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 25 BD: 152 ..-.: &x->wait#2 ->&p->pi_lock FD: 25 BD: 95 ....: &rnp->exp_wq[2] ->&p->pi_lock FD: 30 BD: 6 ....: tasks_rcu_exit_srcu_srcu_usage.lock ->&obj_hash[i].lock FD: 1 BD: 17 ....: &ACCESS_PRIVATE(sdp, lock) FD: 1 BD: 2 ....: tasks_rcu_exit_srcu FD: 37 BD: 2 +.+.: (work_completion)(&(&ssp->srcu_sup->work)->work) ->&ssp->srcu_sup->srcu_gp_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ssp->srcu_sup->srcu_cb_mutex ->remove_cache_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&base->lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 FD: 34 BD: 3 +.+.: &ssp->srcu_sup->srcu_gp_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&rq->__lock ->&ssp->srcu_sup->srcu_cb_mutex ->remove_cache_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 17 ....: &x->wait#3 ->&p->pi_lock FD: 33 BD: 4 +.+.: &ssp->srcu_sup->srcu_cb_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->remove_cache_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 29 BD: 2 +.+.: (work_completion)(&sdp->work) ->&ACCESS_PRIVATE(sdp, lock) ->&obj_hash[i].lock ->&x->wait#3 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 2 ....: kernel/rcu/tasks.h:152 FD: 25 BD: 94 ....: &rnp->exp_wq[3] ->&p->pi_lock FD: 12 BD: 88 ..-.: rcu_tasks_trace__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 87 ....: &ACCESS_PRIVATE(rtpcp, lock) FD: 1 BD: 1 +.+.: (memory_chain).rwsem FD: 51 BD: 89 +.+.: smpboot_threads_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->hrtimer_bases.lock FD: 25 BD: 474 ..-.: &rcu_state.gp_wq ->&p->pi_lock FD: 24 BD: 89 -.-.: &stop_pi_lock ->&rq->__lock FD: 1 BD: 89 -.-.: &stopper->lock FD: 1 BD: 2 +.+.: (module_notify_list).rwsem FD: 1 BD: 1 +.+.: ddebug_lock FD: 2 BD: 1 +.+.: cci_probing ->devtree_lock FD: 1 BD: 1 +.+.: ptlock_ptr(ptdesc) FD: 229 BD: 1 +.+.: watchdog_mutex ->cpu_hotplug_lock FD: 1012 BD: 1 +.+.: (wq_completion)events ->(work_completion)(&sscs.work) ->rdist_memreserve_cpuhp_cleanup_work ->(shepherd).work ->(work_completion)(&(&group->avgs_work)->work) ->(work_completion)(&rfkill_global_led_trigger_work) ->timer_update_work ->pcpu_balance_work ->(work_completion)(&p->wq) ->(debug_obj_work).work ->(work_completion)(&helper->damage_work) ->(work_completion)(&rfkill->sync_work) ->(linkwatch_work).work ->(work_completion)(&w->work) ->(work_completion)(&gadget->work) ->kernfs_notify_work ->async_lookup_work ->autoload_work ->(work_completion)(&barr->work) ->drain_vmap_work ->netstamp_work ->reg_work ->(work_completion)(&fw_work->work) ->(work_completion)(&s->destroy_work) ->(work_completion)(&(&krcp->monitor_work)->work) ->(work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&ht->run_work) ->(work_completion)(&aux->work) ->(work_completion)(&w->work)#2 ->(deferred_probe_timeout_work).work ->(work_completion)(&sbi->s_sb_upd_work) ->(work_completion)(&cgrp->bpf.release_work) ->(work_completion)(&w->w) ->deferred_process_work ->(work_completion)(&data->fib_event_work) ->(work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->(work_completion)(&(&hwstats->traffic_dw)->work) ->(regulator_init_complete_work).work ->(work_completion)(&rdev->wiphy_work) ->wireless_nlevent_work ->(work_completion)(&(&conn->info_timer)->work) ->free_ipc_work ->&rq->__lock ->(work_completion)(&(&devlink->rwork)->work) ->fqdir_free_work ->(work_completion)(&uhid->worker) ->(ima_keys_delayed_work).work ->(work_completion)(&ns->work) ->(work_completion)(&aux->work)#2 FD: 16 BD: 2 +.+.: (work_completion)(&sscs.work) ->&x->wait#4 ->&obj_hash[i].lock ->hrtimer_bases.lock ->&x->wait#5 FD: 1 BD: 90 -.-.: &x->wait#4 FD: 1 BD: 88 ....: &x->wait#5 FD: 2 BD: 129 +.+.: &newf->file_lock ->&newf->resize_wait FD: 1 BD: 1 ....: &p->vtime.seqcount FD: 40 BD: 87 +.+.: mem_hotplug_lock ->mem_hotplug_lock.rss.gp_wait.lock FD: 3 BD: 88 ..-.: mem_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 1 BD: 87 ....: mem_hotplug_lock.waiters.lock FD: 230 BD: 3 +.+.: cpu_add_remove_lock ->cpu_hotplug_lock ->cpu_hotplug_lock.waiters.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->cpuset_hotplug_work FD: 3 BD: 87 ..-.: cpu_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 25 BD: 87 ....: cpu_hotplug_lock.waiters.lock ->&p->pi_lock FD: 1 BD: 4 +.+.: cpuset_hotplug_work FD: 1 BD: 88 +.+.: pcp_batch_high_lock FD: 25 BD: 94 ....: (cpu_running).wait.lock ->&p->pi_lock FD: 25 BD: 197 +.-.: (&timer.timer) ->&p->pi_lock FD: 1 BD: 1 ....: rcu_callback FD: 25 BD: 94 ....: &x->wait#6 ->&p->pi_lock FD: 1 BD: 3518 -.-.: &cfs_rq->removed.lock FD: 1 BD: 93 ....: &x->wait#7 FD: 15 BD: 3515 -.-.: &rt_b->rt_runtime_lock ->&rt_rq->rt_runtime_lock ->tk_core.seq.seqcount ->hrtimer_bases.lock FD: 1 BD: 3516 -.-.: &rt_rq->rt_runtime_lock FD: 27 BD: 87 +.+.: stop_cpus_mutex ->&stopper->lock ->&stop_pi_lock ->&rq->__lock ->&x->wait#8 FD: 1 BD: 89 ....: &x->wait#8 FD: 229 BD: 2 +.+.: rdist_memreserve_cpuhp_cleanup_work ->cpu_hotplug_lock FD: 58 BD: 1 +.+.: sched_domains_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->pcpu_lock FD: 1 BD: 3515 ....: &cp->lock FD: 77 BD: 1 +.+.: &type->s_umount_key#5/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock FD: 37 BD: 113 +.+.: &sb->s_type->i_lock_key#5 ->&dentry->d_lock FD: 25 BD: 1 ....: (setup_done).wait.lock ->&p->pi_lock FD: 75 BD: 26 ++++: namespace_sem ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->rename_lock ->&obj_hash[i].lock ->&rq->__lock ->&____s->seqcount#2 ->&n->list_lock ->stock_lock FD: 1 BD: 123 +.+.: &____s->seqcount#3 FD: 61 BD: 1 +.+.: &type->s_umount_key#6 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->&lru->node[i].lock ->&sbinfo->stat_lock ->&obj_hash[i].lock FD: 25 BD: 413 +.+.: &lru->node[i].lock FD: 82 BD: 7 ++++: &sb->s_type->i_mutex_key ->namespace_sem ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->tomoyo_ss ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#2 ->&wb->list_lock ->&c->lock ->&____s->seqcount ->&rq->__lock FD: 38 BD: 23 +.+.: rename_lock ->rename_lock.seqcount FD: 37 BD: 137 +.+.: rename_lock.seqcount ->&dentry->d_lock ->&dentry->d_lock/2 FD: 1 BD: 122 ....: &new_ns->poll FD: 38 BD: 124 +.+.: &fs->lock ->&____s->seqcount#3 ->&dentry->d_lock FD: 1 BD: 86 +.+.: req_lock FD: 95 BD: 1 +.+.: of_mutex ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem FD: 1 BD: 188 ....: &x->wait#9 FD: 1 BD: 211 +.+.: &k->list_lock FD: 24 BD: 188 ++++: bus_type_sem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 32 BD: 260 -.-.: &dev->power.lock ->&dev->power.wait_queue ->hrtimer_bases.lock ->&dev->power.lock/1 FD: 24 BD: 186 +.+.: dpm_list_mtx ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 61 BD: 188 +.+.: uevent_sock_mutex ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&rq->__lock ->&cfs_rq->removed.lock ->quarantine_lock ->rcu_node_0 ->rlock-AF_NETLINK ->&n->list_lock ->remove_cache_srcu ->uevent_sock_mutex.wait_lock ->&____s->seqcount#2 ->mmu_notifier_invalidate_range_start ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock FD: 1 BD: 133 ....: running_helpers_waitq.lock FD: 1 BD: 200 +.+.: sysfs_symlink_target_lock FD: 2 BD: 254 +.+.: &k->k_lock ->klist_remove_lock FD: 1 BD: 1 ....: &dev->mutex FD: 1 BD: 1 +.+.: subsys mutex FD: 2 BD: 1 +.+.: memory_blocks.xa_lock ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#2 FD: 1 BD: 1 +.+.: subsys mutex#3 FD: 57 BD: 125 +.+.: dev_pm_qos_mtx ->fs_reclaim ->pool_lock#2 ->&dev->power.lock ->pm_qos_lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 126 ....: pm_qos_lock FD: 101 BD: 123 +.+.: dev_pm_qos_sysfs_mtx ->dev_pm_qos_mtx ->&root->kernfs_rwsem ->fs_reclaim ->pool_lock#2 ->lock ->&____s->seqcount ->&c->lock ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->dev_pm_qos_sysfs_mtx.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 51 BD: 13 +.+.: register_lock ->proc_subdir_lock ->fs_reclaim ->pool_lock#2 ->proc_inum_ida.xa_lock ->&c->lock ->&____s->seqcount FD: 1 BD: 1 +.+.: (cpufreq_policy_notifier_list).rwsem FD: 1 BD: 2 +.+.: (pm_chain_head).rwsem FD: 1 BD: 1 +.+.: cpufreq_governor_mutex FD: 40 BD: 2 +.+.: (work_completion)(&rew->rew_work) ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&rcu_state.expedited_wq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->pool_lock#2 ->&cfs_rq->removed.lock ->rcu_state.exp_wake_mutex.wait_lock ->&p->pi_lock FD: 30 BD: 1 ..-.: &(&kfence_timer)->timer FD: 25 BD: 774 -.-.: &rcu_state.expedited_wq ->&p->pi_lock FD: 1 BD: 2 ++++: binfmt_lock FD: 1 BD: 75 +.+.: pin_fs_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#7/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&____s->seqcount ->&c->lock ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#6 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 3 +.+.: &sb->s_type->i_lock_key#6 ->&dentry->d_lock FD: 63 BD: 1 +.+.: &sb->s_type->i_mutex_key#2 ->&sb->s_type->i_lock_key#6 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount FD: 25 BD: 414 ....: &wq ->&p->pi_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#8/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 73 +.+.: &sb->s_type->i_lock_key#7 ->&dentry->d_lock FD: 88 BD: 71 +.+.: &sb->s_type->i_mutex_key#3 ->&sb->s_type->i_lock_key#7 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->rcu_node_0 ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&n->list_lock ->(console_sem).lock ->&rcu_state.expedited_wq ->pin_fs_lock ->mount_lock ->&fsnotify_mark_srcu ->&xa->xa_lock#9 ->&____s->seqcount#2 ->&rcu_state.gp_wq ->remove_cache_srcu ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 22 +.+.: chrdevs_lock FD: 872 BD: 2 ++++: cb_lock ->genl_mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->rlock-AF_NETLINK ->rtnl_mutex ->&obj_hash[i].lock ->&____s->seqcount ->&rdev->wiphy.mtx ->nlk_cb_mutex-GENERIC ->&n->list_lock ->&xa->xa_lock#16 ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&____s->seqcount#2 ->&lock->wait_lock ->genl_mutex.wait_lock ->remove_cache_srcu ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock FD: 859 BD: 3 +.+.: genl_mutex ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->nl_table_wait.lock ->rlock-AF_NETLINK ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&zone->lock ->hwsim_radio_lock ->&x->wait#9 ->batched_entropy_u32.lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#53 ->device_links_lock ->&k->k_lock ->deferred_probe_mutex ->pcpu_alloc_mutex ->cpu_hotplug_lock ->wq_pool_mutex ->crngs.lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->(inetaddr_chain).rwsem ->inet6addr_chain.lock ->&____s->seqcount#2 ->genl_mutex.wait_lock ->quarantine_lock ->remove_cache_srcu ->&sem->wait_lock ->key ->pcpu_lock ->percpu_counters_lock ->pcpu_alloc_mutex.wait_lock ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock FD: 1 BD: 4 +.+.: subsys mutex#4 FD: 4 BD: 5 ....: async_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 357 BD: 2 +.+.: (work_completion)(&entry->work) ->tk_core.seq.seqcount ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&k->list_lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->&c->lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->&rq->__lock ->async_lock ->async_done.lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->sb_writers#2 ->&sb->s_type->i_lock_key#2 FD: 1 BD: 21 .+.+: device_links_srcu FD: 3 BD: 20 +.+.: fwnode_link_lock ->&k->k_lock FD: 1 BD: 3 +.+.: regulator_list_mutex FD: 28 BD: 125 +.+.: device_links_lock ->&k->list_lock ->&k->k_lock ->&rq->__lock ->device_links_lock.wait_lock FD: 1 BD: 22 ....: &dev->devres_lock FD: 24 BD: 20 +.+.: pinctrl_list_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 1 BD: 20 +.+.: pinctrl_maps_mutex FD: 1 BD: 4 +.+.: regulator_nesting_mutex FD: 2 BD: 1 +.+.: regulator_ww_class_mutex ->regulator_nesting_mutex FD: 99 BD: 92 +.+.: gdp_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->kobj_ns_type_lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&sem->wait_lock ->&p->pi_lock ->sysfs_symlink_target_lock ->kernfs_idr_lock ->&obj_hash[i].lock ->gdp_mutex.wait_lock FD: 1 BD: 1 +.+.: (reboot_notifier_list).rwsem FD: 3 BD: 3 +.+.: subsys mutex#5 ->&k->k_lock FD: 24 BD: 126 +.+.: deferred_probe_mutex ->&rq->__lock FD: 1 BD: 20 ....: probe_waitqueue.lock FD: 25 BD: 3 ....: async_done.lock ->&p->pi_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#9/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 49 +.+.: &sb->s_type->i_lock_key#8 ->&dentry->d_lock ->&p->pi_lock FD: 33 BD: 7 +.+.: vmap_purge_lock ->purge_vmap_area_lock ->free_vmap_area_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 4 +.+.: &fp->aux->used_maps_mutex FD: 1 BD: 1 +.+.: proto_list_mutex FD: 1 BD: 1 +.+.: targets_mutex FD: 26 BD: 404 ...-: nl_table_lock ->pool_lock#2 ->nl_table_wait.lock ->&obj_hash[i].lock ->&c->lock FD: 25 BD: 405 ..-.: nl_table_wait.lock ->&p->pi_lock FD: 1 BD: 1 +.+.: net_family_lock FD: 2 BD: 5 ....: net_generic_ids.xa_lock ->pool_lock#2 FD: 6 BD: 52 ..-.: &dir->lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock FD: 36 BD: 5 +.+.: k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK FD: 1 BD: 6 +...: k-slock-AF_NETLINK FD: 2 BD: 98 ..-.: rhashtable_bucket ->rhashtable_bucket/1 FD: 1 BD: 47 ....: &cma->lock FD: 43 BD: 47 +.+.: cma_mutex ->&zone->lock ->&obj_hash[i].lock ->lock#2 FD: 27 BD: 1 +.+.: pcpu_drain_mutex ->&pcp->lock FD: 38 BD: 49 +.+.: lock#2 ->&obj_hash[i].lock ->(work_completion)(work) ->&x->wait#10 ->&rq->__lock FD: 1 BD: 1 +.+.: &pool->lock#2 FD: 1 BD: 90 ....: freezer_lock FD: 1 BD: 1 ....: audit_backlog_wait.lock FD: 1 BD: 1 ....: &list->lock FD: 1 BD: 1 ....: kauditd_wait.lock FD: 1 BD: 1 ....: printk_ratelimit_state.lock FD: 3 BD: 2 +.+.: lock#3 ->&zone->lock FD: 53 BD: 1 +.+.: khugepaged_mutex ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->lock#3 ->pcp_batch_high_lock FD: 125 BD: 9 ++++: &(&priv->bus_notifier)->rwsem ->&device->physical_node_lock ->iommu_probe_device_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->i2c_dev_list_lock ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->gdp_mutex ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#61 ->&sem->wait_lock ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 3 BD: 6 +.+.: subsys mutex#6 ->&k->k_lock FD: 4 BD: 1 +.+.: subsys mutex#7 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 +.+.: regmap_debugfs_early_lock FD: 1 BD: 1 +.+.: (acpi_reconfig_chain).rwsem FD: 1 BD: 1 +.+.: __i2c_board_lock FD: 1 BD: 707 -.-.: quarantine_lock FD: 52 BD: 1 +.+.: core_lock ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 FD: 35 BD: 561 .+.+: remove_cache_srcu ->quarantine_lock ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&____s->seqcount ->rcu_node_0 ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->&base->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 2 BD: 1 +.+.: thermal_governor_lock ->thermal_list_lock FD: 1 BD: 2 +.+.: thermal_list_lock FD: 32 BD: 1 +.+.: cpuidle_lock ->&obj_hash[i].lock ->(console_sem).lock FD: 52 BD: 1 +.+.: k-sk_lock-AF_QIPCRTR ->k-slock-AF_QIPCRTR ->fs_reclaim ->qrtr_ports.xa_lock ->pool_lock#2 ->qrtr_node_lock ->&obj_hash[i].lock FD: 1 BD: 2 +...: k-slock-AF_QIPCRTR FD: 1 BD: 2 +.+.: qrtr_ports.xa_lock FD: 1 BD: 2 +.+.: qrtr_node_lock FD: 51 BD: 94 ++++: (crypto_chain).rwsem ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 293 BD: 2 +.+.: tty_mutex ->(console_sem).lock ->console_lock ->fs_reclaim ->pool_lock#2 ->tty_ldiscs_lock ->&obj_hash[i].lock ->&k->list_lock ->&k->k_lock ->&tty->legacy_mutex ->stock_lock ->&c->lock ->&rq->__lock ->tty_mutex.wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->&cfs_rq->removed.lock FD: 230 BD: 1 +.+.: iova_cache_mutex ->cpu_hotplug_lock ->slab_mutex FD: 3 BD: 1 +.+.: subsys mutex#8 ->&k->k_lock FD: 1 BD: 1 ..-.: uidhash_lock FD: 8 BD: 2 +.+.: (work_completion)(&s->destroy_work) ->&obj_hash[i].lock ->pool_lock#2 ->&rsp->gp_wait ->pcpu_lock FD: 52 BD: 1 +.+.: (work_completion)(&eval_map_work) ->trace_event_sem FD: 1 BD: 1 ....: oom_reaper_wait.lock FD: 1 BD: 1 +.+.: subsys mutex#9 FD: 1 BD: 1 ....: &pgdat->kcompactd_wait FD: 1 BD: 1 ....: hugetlb_lock FD: 116 BD: 1 +.+.: memory_tier_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#10 FD: 1 BD: 2 +.+.: subsys mutex#10 FD: 1 BD: 1 +.+.: ksm_thread_mutex FD: 1 BD: 1 ....: ksm_thread_wait.lock FD: 1 BD: 1 +.+.: khugepaged_mm_lock FD: 25 BD: 8 ....: khugepaged_wait.lock ->&p->pi_lock FD: 1 BD: 2 +.+.: damon_ops_lock FD: 52 BD: 93 ++++: crypto_alg_sem ->(crypto_chain).rwsem FD: 103 BD: 8 +.+.: bio_slab_lock ->fs_reclaim ->pool_lock#2 ->slab_mutex ->bio_slabs.xa_lock FD: 2 BD: 9 +.+.: bio_slabs.xa_lock ->pool_lock#2 FD: 50 BD: 1 +.+.: major_names_lock ->fs_reclaim ->pool_lock#2 ->major_names_spinlock FD: 1 BD: 2 +.+.: major_names_spinlock FD: 14 BD: 1 +.-.: (&rtpcp->lazy_timer) ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->rcu_tasks__percpu.cbs_pcpu_lock FD: 51 BD: 1 +.+.: &pgdat->kswapd_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock FD: 30 BD: 1 ..-.: drivers/char/random.c:251 FD: 13 BD: 2 +.+.: (next_reseed).work ->&obj_hash[i].lock ->&base->lock ->input_pool.lock ->base_crng.lock FD: 30 BD: 1 ..-.: mm/vmstat.c:2022 FD: 229 BD: 2 +.+.: (shepherd).work ->cpu_hotplug_lock ->&obj_hash[i].lock ->&base->lock FD: 34 BD: 50 +.+.: (wq_completion)mm_percpu_wq ->(work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->(work_completion)(work) ->(work_completion)(&barr->work) FD: 27 BD: 51 +.+.: (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->&obj_hash[i].lock ->&base->lock ->&pcp->lock FD: 3 BD: 6 +.+.: subsys mutex#11 ->&k->k_lock FD: 1 BD: 1 -...: shrink_qlist.lock FD: 30 BD: 5 ....: remove_cache_srcu_srcu_usage.lock ->&obj_hash[i].lock FD: 30 BD: 1 ..-.: &(&ssp->srcu_sup->work)->timer FD: 33 BD: 87 +.+.: flush_lock ->&obj_hash[i].lock ->(work_completion)(&sfw->work) ->&x->wait#10 ->&rq->__lock FD: 9 BD: 89 +.+.: (work_completion)(&sfw->work) ->&c->lock ->&n->list_lock ->&obj_hash[i].lock FD: 28 BD: 88 +.+.: (wq_completion)slub_flushwq ->(work_completion)(&sfw->work) ->(work_completion)(&barr->work) FD: 25 BD: 659 ....: &x->wait#10 ->&p->pi_lock FD: 26 BD: 94 +.+.: (work_completion)(&barr->work) ->&x->wait#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: prepare_lock FD: 89 BD: 1 +.+.: clk_debug_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 1 BD: 1 +.+.: clocks_mutex FD: 345 BD: 1 +.+.: acpi_scan_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#9 ->acpi_device_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#12 ->uevent_sock_mutex ->running_helpers_waitq.lock ->*(&acpi_gbl_reference_count_lock) ->&rq->__lock ->&n->list_lock ->quarantine_lock ->acpi_ioremap_lock ->&device->physical_node_lock ->irq_domain_mutex ->&domain->mutex ->resource_lock ->&(&priv->bus_notifier)->rwsem ->fwnode_link_lock ->device_links_srcu ->acpi_pm_notifier_install_lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->subsys mutex#4 ->(console_sem).lock ->io_range_mutex ->pci_bus_sem ->gdp_mutex ->subsys mutex#17 ->acpi_hp_context_lock ->bridge_mutex ->pci_lock ->pci_acpi_companion_lookup_sem ->pci_slot_mutex ->resource_alignment_lock ->iort_msi_chip_lock ->subsys mutex#18 ->devtree_lock ->pci_rescan_remove_lock ->acpi_link_lock ->acpi_dep_list_lock ->power_resource_list_lock FD: 51 BD: 2 +.+.: acpi_device_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 1 BD: 3 ....: &xa->xa_lock#2 FD: 1 BD: 2 +.+.: subsys mutex#12 FD: 96 BD: 33 +.+.: &device->physical_node_lock ->sysfs_symlink_target_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock FD: 119 BD: 2 +.+.: acpi_pm_notifier_install_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->*(&acpi_gbl_reference_count_lock) ->acpi_pm_notifier_lock ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 116 BD: 3 +.+.: acpi_pm_notifier_lock ->fs_reclaim ->pool_lock#2 ->wakeup_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#13 ->events_lock FD: 1 BD: 5 ....: wakeup_ida.xa_lock FD: 3 BD: 5 +.+.: subsys mutex#13 ->&k->k_lock FD: 1 BD: 5 ....: events_lock FD: 1 BD: 2 +.+.: acpi_wakeup_lock FD: 163 BD: 2 +.+.: port_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#14 ->&xa->xa_lock#3 ->&port->mutex FD: 1 BD: 3 +.+.: subsys mutex#14 FD: 1 BD: 3 ....: &xa->xa_lock#3 FD: 1 BD: 261 ....: &dev->power.wait_queue FD: 79 BD: 1 +.+.: (wq_completion)pm ->(work_completion)(&dev->power.work) ->(work_completion)(&hcd->wakeup_work) FD: 73 BD: 2 +.+.: (work_completion)(&dev->power.work) ->&dev->power.lock ->&hub->irq_urb_lock ->(&hub->irq_urb_retry) ->&obj_hash[i].lock ->&base->lock ->hcd_urb_unlink_lock ->hcd_root_hub_lock ->usb_kill_urb_queue.lock ->&rq->__lock ->(work_completion)(&hub->tt.clear_work) ->&dum_hcd->dum->lock ->device_state_lock ->hcd_urb_list_lock ->&vhci_hcd->vhci->lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->fs_reclaim ->&x->wait#19 ->(&timer.timer) ->&c->lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&port_lock_key FD: 154 BD: 9 +.+.: &port->mutex ->fs_reclaim ->pool_lock#2 ->(console_sem).lock ->&port_lock_key ->console_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->ctrl_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&dev->power.lock ->&____s->seqcount ->&k->list_lock ->&c->lock ->lock ->&root->kernfs_rwsem ->&device->physical_node_lock ->semaphore->lock ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#15 ->*(&acpi_gbl_reference_count_lock) ->&n->list_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->gdp_mutex ->bus_type_sem ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->subsys mutex#16 ->chrdevs_lock ->&cfs_rq->removed.lock ->quarantine_lock ->&desc->request_mutex ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock FD: 35 BD: 218 ....: &port_lock_key ->&dev->power.lock ->&port->lock ->&tty->write_wait FD: 24 BD: 11 +.+.: syslog_lock ->&rq->__lock FD: 37 BD: 214 ....: console_owner ->&port_lock_key ->console_owner_lock FD: 30 BD: 1 ..-.: &(&group->avgs_work)->timer FD: 30 BD: 1 ..-.: &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer FD: 30 BD: 1 ..-.: mm/memcontrol.c:679 FD: 25 BD: 2 +.+.: (work_completion)(&(&group->avgs_work)->work) ->&group->avgs_lock FD: 24 BD: 3 +.+.: &group->avgs_lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 25 BD: 2 +.+.: (stats_flush_dwork).work ->cgroup_rstat_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 2 BD: 18 ....: cgroup_rstat_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 3537 ..-.: per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 10 ....: ctrl_ida.xa_lock FD: 1 BD: 10 +.+.: subsys mutex#15 FD: 1 BD: 255 +.+.: klist_remove_lock FD: 25 BD: 86 ....: &x->wait#11 ->&p->pi_lock FD: 100 BD: 1 .+.+: sb_writers ->mount_lock ->&type->i_mutex_dir_key/1 ->&sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#5 ->&wb->list_lock ->&type->i_mutex_dir_key#2 ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->&s->s_inode_list_lock ->&obj_hash[i].lock ->&sbinfo->stat_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu FD: 92 BD: 2 +.+.: &type->i_mutex_dir_key/1 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&c->lock ->&____s->seqcount ->&xattrs->lock ->&simple_offset_xa_lock ->smack_known_lock ->&sb->s_type->i_mutex_key#4 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->tomoyo_ss ->&u->bindlock ->&n->list_lock ->&sem->wait_lock ->&rq->__lock ->remove_cache_srcu ->smack_known_lock.wait_lock ->&p->pi_lock ->rcu_node_0 ->&fsnotify_mark_srcu ->&xa->xa_lock#9 ->&type->i_mutex_dir_key#2 ->&____s->seqcount#2 FD: 1 BD: 10 ++++: &xattrs->lock FD: 13 BD: 10 +.+.: &simple_offset_xa_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 37 BD: 95 +.+.: smack_known_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->smack_known_lock.wait_lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&c->lock ->&base->lock FD: 28 BD: 3 +.+.: &sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->tomoyo_ss ->&xattrs->lock ->&rq->__lock ->&simple_offset_xa_lock ->&dentry->d_lock FD: 3 BD: 10 +.+.: subsys mutex#16 ->&k->k_lock FD: 1 BD: 2 +.+.: io_range_mutex FD: 1 BD: 2 ++++: pci_bus_sem FD: 3 BD: 2 +.+.: subsys mutex#17 ->&k->k_lock FD: 49 BD: 2 +.+.: acpi_hp_context_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 2 +.+.: bridge_mutex FD: 1 BD: 13 ....: pci_lock FD: 1 BD: 2 .+.+: pci_acpi_companion_lookup_sem FD: 1 BD: 2 +.+.: pci_slot_mutex FD: 1 BD: 2 +.+.: resource_alignment_lock FD: 1 BD: 261 ....: &dev->power.lock/1 FD: 1 BD: 2 +.+.: subsys mutex#18 FD: 36 BD: 2 +.+.: pci_rescan_remove_lock FD: 64 BD: 4 +.+.: acpi_link_lock ->fs_reclaim ->pool_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->*(&acpi_gbl_reference_count_lock) ->(console_sem).lock ->&c->lock ->&____s->seqcount FD: 1 BD: 2 +.+.: acpi_dep_list_lock FD: 1 BD: 2 +.+.: power_resource_list_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#10/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#9 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#9 ->&dentry->d_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#11/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#10 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#10 ->&dentry->d_lock FD: 161 BD: 77 ++++: &mm->mmap_lock ->reservation_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&vma->vm_lock->lock ->&____s->seqcount ->&mm->page_table_lock ->ptlock_ptr(ptdesc)#2 ->&c->lock ->&anon_vma->rwsem ->mmu_notifier_invalidate_range_start ->lock#4 ->lock#5 ->&obj_hash[i].lock ->&mapping->i_mmap_rwsem ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&p->alloc_lock ->&lruvec->lru_lock ->tk_core.seq.seqcount ->rcu_node_0 ->&rq->__lock ->&mm->mmap_lock/1 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->key ->pcpu_lock ->percpu_counters_lock ->&n->list_lock ->remove_cache_srcu ->&sem->wait_lock ->&p->pi_lock ->pool_lock ->&cfs_rq->removed.lock ->&folio_wait_table[i] ->quarantine_lock ->&sb->s_type->i_lock_key ->&kcov->lock ->&rcu_state.expedited_wq ->stock_lock ->&____s->seqcount#2 ->&base->lock ->&xa->xa_lock#9 ->&info->lock FD: 74 BD: 88 +.+.: reservation_ww_class_acquire ->reservation_ww_class_mutex FD: 73 BD: 89 +.+.: reservation_ww_class_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&xa->xa_lock#9 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock FD: 46 BD: 707 ++++: &mapping->i_mmap_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&anon_vma->rwsem ->&rq->__lock ->&sem->wait_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock FD: 1 BD: 729 +.+.: dma_fence_map FD: 1 BD: 756 ....: key FD: 1 BD: 1 +.+.: attribute_container_mutex FD: 24 BD: 14 +.+.: triggers_list_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 14 .+.+: leds_list_lock ->&rq->__lock FD: 136 BD: 4 ++++: (usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&____s->seqcount ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#57 ->mon_lock ->usbfs_mutex FD: 1 BD: 1 +.+.: rc_map_lock FD: 1 BD: 1 +.+.: subsys mutex#19 FD: 2 BD: 10 +.+.: iommu_probe_device_lock ->iommu_device_lock FD: 1 BD: 11 +.+.: iommu_device_lock FD: 1 BD: 7 ....: (efi_runtime_lock).lock FD: 25 BD: 9 ....: &x->wait#12 ->&p->pi_lock FD: 28 BD: 1 +.+.: (wq_completion)efi_rts_wq ->(work_completion)(&efi_rts_work.work) FD: 27 BD: 2 +.+.: (work_completion)(&efi_rts_work.work) ->cpu_asid_lock ->efi_rt_lock ->&x->wait#12 ->&rq->__lock FD: 1 BD: 3515 ....: cpu_asid_lock FD: 1 BD: 3 +.+.: efi_rt_lock FD: 1 BD: 6 ....: (efivars_lock).lock FD: 1 BD: 1 +.+.: devfreq_list_lock FD: 1 BD: 5 +.+.: &entry->access FD: 51 BD: 5 +.+.: info_mutex ->proc_subdir_lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->proc_inum_ida.xa_lock ->&rq->__lock ->&____s->seqcount#2 FD: 1 BD: 93 +.+.: kobj_ns_type_lock FD: 13 BD: 55 +.+.: &xa->xa_lock#4 ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 26 BD: 49 +.+.: subsys mutex#20 ->&k->k_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 7 BD: 243 ..-.: &dir->lock#2 ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 37 BD: 55 +.+.: dev_hotplug_mutex ->&dev->power.lock ->&k->k_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 14 BD: 54 ++++: dev_base_lock ->&xa->xa_lock#4 FD: 1 BD: 1 +.+.: qdisc_mod_lock FD: 18 BD: 1 ++++: bt_proto_lock ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->chan_list_lock ->l2cap_sk_list.lock ->&sk->sk_peer_lock ->hci_sk_list.lock ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 61 BD: 50 +.+.: hci_cb_list_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->chan_list_lock ->&conn->ident_lock ->&base->lock ->&c->lock ->&list->lock#8 ->&conn->chan_lock ->&____s->seqcount ->pool_lock ->(work_completion)(&(&conn->id_addr_timer)->work) ->&x->wait#2 ->&rq->__lock ->(work_completion)(&(&conn->info_timer)->work) ->hci_cb_list_lock.wait_lock ->&n->list_lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 FD: 1 BD: 1 +.+.: mgmt_chan_list_lock FD: 1 BD: 83 ....: &list->lock#2 FD: 49 BD: 47 +.+.: rate_ctrl_mutex ->fs_reclaim ->pool_lock#2 FD: 2 BD: 1 +.+.: netlbl_domhsh_lock ->pool_lock#2 FD: 1 BD: 45 +.+.: netlbl_unlhsh_lock FD: 340 BD: 1 +.+.: misc_mtx ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&____s->seqcount ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#21 ->misc_minors_ida.xa_lock ->&cfs_rq->removed.lock ->&base->lock ->&dir->lock ->pool_lock ->rfkill_global_mutex ->nfc_index_ida.xa_lock ->&____s->seqcount#2 ->pcpu_alloc_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&x->wait ->hrtimer_bases.lock ->wq_pool_mutex ->nfc_devlist_mutex ->&k->k_lock ->llcp_devices_lock ->&n->list_lock ->nl_table_lock ->nl_table_wait.lock ->misc_mtx.wait_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->nfc_devlist_mutex.wait_lock ->wq_pool_mutex.wait_lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pcpu_alloc_mutex.wait_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&meta->lock ->(console_sem).lock ->&wq->mutex ->(wq_completion)nfc2_nci_tx_wq#1631 ->(wq_completion)nfc2_nci_rx_wq#1631 ->wq_mayday_lock ->(wq_completion)nfc2_nci_cmd_wq#1630 ->(wq_completion)nfc2_nci_rx_wq#1683 ->(wq_completion)nfc2_nci_cmd_wq#1682 ->(wq_completion)nfc2_nci_cmd_wq#1684 ->console_owner_lock ->console_owner ->(wq_completion)nfc2_nci_tx_wq#1698 ->(wq_completion)nfc2_nci_rx_wq#1699 ->(wq_completion)nfc2_nci_cmd_wq#1699 ->(wq_completion)nfc3_nci_rx_wq#1488 ->(wq_completion)nfc3_nci_cmd_wq#1488 ->(wq_completion)nfc2_nci_tx_wq#1716 ->(wq_completion)nfc2_nci_rx_wq#1717 ->(wq_completion)nfc2_nci_cmd_wq#1717 ->(wq_completion)nfc2_nci_rx_wq#1760 ->(wq_completion)nfc2_nci_cmd_wq#1760 ->(wq_completion)nfc2_nci_rx_wq#1762 ->(wq_completion)nfc2_nci_cmd_wq#1762 ->(wq_completion)nfc2_nci_rx_wq#1773 ->(wq_completion)nfc2_nci_cmd_wq#1773 ->(wq_completion)nfc3_nci_cmd_wq#1549 ->(wq_completion)nfc5_nci_cmd_wq#304 ->(wq_completion)nfc8_nci_rx_wq#25 ->(wq_completion)nfc8_nci_cmd_wq#25 ->(wq_completion)nfc8_nci_tx_wq#25 ->(wq_completion)nfc8_nci_rx_wq#26 ->(wq_completion)nfc8_nci_cmd_wq#26 ->(wq_completion)nfc8_nci_rx_wq#27 ->(wq_completion)nfc8_nci_cmd_wq#27 ->(wq_completion)nfc2_nci_cmd_wq#1890 ->(wq_completion)nfc2_nci_cmd_wq#1897 ->(wq_completion)nfc2_nci_rx_wq#1902 ->(wq_completion)nfc2_nci_cmd_wq#1904 ->(wq_completion)nfc4_nci_tx_wq#470 ->(wq_completion)nfc4_nci_rx_wq#470 ->(wq_completion)nfc4_nci_cmd_wq#470 ->(wq_completion)nfc2_nci_tx_wq#1917 ->(wq_completion)nfc2_nci_rx_wq#1922 ->(wq_completion)nfc2_nci_cmd_wq#1924 ->(wq_completion)nfc6_nci_rx_wq#95 ->(wq_completion)nfc6_nci_cmd_wq#95 ->(wq_completion)nfc3_nci_tx_wq#1659 ->(wq_completion)nfc3_nci_rx_wq#1659 ->(wq_completion)nfc3_nci_cmd_wq#1660 ->(wq_completion)nfc2_nci_cmd_wq#1945 ->(wq_completion)nfc2_nci_cmd_wq#1946 ->(wq_completion)nfc15_nci_tx_wq#3 ->(wq_completion)nfc15_nci_rx_wq#3 ->(wq_completion)nfc15_nci_cmd_wq#3 ->(wq_completion)nfc9_nci_cmd_wq#23 ->(wq_completion)nfc4_nci_cmd_wq#498 ->(wq_completion)nfc4_nci_cmd_wq#499 ->(wq_completion)nfc4_nci_rx_wq#515 ->(wq_completion)nfc4_nci_cmd_wq#517 ->(wq_completion)nfc5_nci_cmd_wq#354 ->(wq_completion)nfc4_nci_rx_wq#523 ->(wq_completion)nfc4_nci_cmd_wq#525 ->(wq_completion)nfc4_nci_rx_wq#524 ->(wq_completion)nfc4_nci_cmd_wq#526 ->(wq_completion)nfc3_nci_cmd_wq#1714 ->(wq_completion)nfc3_nci_rx_wq#1713 ->(wq_completion)nfc3_nci_cmd_wq#1715 ->(wq_completion)nfc4_nci_tx_wq#528 ->(wq_completion)nfc4_nci_rx_wq#531 ->(wq_completion)nfc4_nci_cmd_wq#533 ->(wq_completion)nfc12_nci_rx_wq#6 ->(wq_completion)nfc12_nci_cmd_wq#6 ->(wq_completion)nfc12_nci_cmd_wq#7 ->(wq_completion)nfc5_nci_cmd_wq#370 ->(wq_completion)nfc5_nci_cmd_wq#371 ->(wq_completion)nfc2_nci_cmd_wq#2020 FD: 3 BD: 2 +.+.: subsys mutex#21 ->&k->k_lock FD: 121 BD: 2 +.+.: (work_completion)(&rfkill_global_led_trigger_work) ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 120 BD: 13 +.+.: rfkill_global_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&____s->seqcount ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->&rfkill->lock ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#40 ->triggers_list_lock ->leds_list_lock ->&rq->__lock ->rfkill_global_mutex.wait_lock ->&n->list_lock ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&data->mtx ->&____s->seqcount#2 ->&sem->wait_lock ->remove_cache_srcu ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 122 BD: 1 +.+.: input_mutex ->input_devices_poll_wait.lock ->fs_reclaim ->pool_lock#2 ->&dev->mutex#2 ->input_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#31 ->&c->lock ->&____s->seqcount FD: 1 BD: 2 ....: input_devices_poll_wait.lock FD: 42 BD: 2 ++++: (netlink_chain).rwsem ->hwsim_radio_lock ->pool_lock#2 ->&obj_hash[i].lock ->reg_indoor_lock ->&rq->__lock ->&c->lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->&____s->seqcount ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: proto_tab_lock FD: 3 BD: 1 ....: random_ready_notifier.lock ->crngs.lock FD: 1 BD: 2 ....: misc_minors_ida.xa_lock FD: 1 BD: 1 +.+.: wtd_deferred_reg_mutex FD: 75 BD: 1 +.+.: &type->s_umount_key#12/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#11 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#11 ->&dentry->d_lock FD: 78 BD: 2 +.+.: &type->s_umount_key#13/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#12 ->&____s->seqcount ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&dentry->d_lock ->&type->s_umount_key#14 FD: 93 BD: 1 +.+.: (work_completion)(&tracerfs_init_work) ->pin_fs_lock ->fs_reclaim ->pool_lock#2 ->sb_lock ->&c->lock ->&____s->seqcount ->&type->s_umount_key#13/1 ->&type->s_umount_key#14 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->&obj_hash[i].lock ->&fsnotify_mark_srcu ->&sb->s_type->i_mutex_key#5 ->event_mutex ->(module_notify_list).rwsem ->trace_types_lock FD: 37 BD: 6 +.+.: &sb->s_type->i_lock_key#12 ->&dentry->d_lock FD: 66 BD: 3 +.+.: &type->s_umount_key#14 ->sb_lock ->list_lrus_mutex ->&xa->xa_lock#5 ->&obj_hash[i].lock ->pool_lock#2 ->shrinker_mutex ->&rsp->gp_wait ->pcpu_lock ->fs_reclaim ->&dentry->d_lock ->&lru->node[i].lock FD: 11 BD: 319 ....: &xa->xa_lock#5 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock FD: 63 BD: 4 +.+.: &sb->s_type->i_mutex_key#5 ->&sb->s_type->i_lock_key#12 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount FD: 4 BD: 10 ..-.: &rsp->gp_wait ->&obj_hash[i].lock ->pool_lock#2 FD: 75 BD: 1 +.+.: &type->s_umount_key#15/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#13 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 51 BD: 119 .+.+: &fsnotify_mark_srcu ->&conn->lock ->fs_reclaim ->pool_lock#2 ->&group->notification_lock ->&group->notification_waitq ->&rq->__lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#13 ->&dentry->d_lock FD: 69 BD: 2 +.+.: event_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 ->trace_event_sem ->trace_types_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#16/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#14 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#14 ->&dentry->d_lock FD: 1 BD: 5 +.+.: eventfs_mutex FD: 75 BD: 1 +.+.: &type->s_umount_key#17/1 ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#15 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#15 ->&dentry->d_lock FD: 230 BD: 2 +.+.: timer_update_work ->timer_keys_mutex FD: 229 BD: 3 +.+.: timer_keys_mutex ->cpu_hotplug_lock FD: 50 BD: 1 +.+.: kclist_lock ->resource_lock ->fs_reclaim ->pool_lock#2 FD: 75 BD: 1 +.+.: &type->s_umount_key#18/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#16 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#16 ->&dentry->d_lock FD: 184 BD: 30 .+.+: tomoyo_ss ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tomoyo_policy_lock ->(console_sem).lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&dentry->d_lock ->tomoyo_log_lock ->tomoyo_log_wait.lock ->&rq->__lock ->file_systems_lock ->fs_reclaim ->quarantine_lock ->&mm->mmap_lock ->key ->pcpu_lock ->percpu_counters_lock ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->mount_lock ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->stock_lock ->&base->lock ->(kmod_concurrent_max).lock ->&x->wait#17 ->running_helpers_waitq.lock ->&fs->lock ->&type->i_mutex_dir_key#3 FD: 49 BD: 1 +.+.: pnp_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#22 FD: 3 BD: 1 +.+.: subsys mutex#23 ->&k->k_lock FD: 3 BD: 1 +.+.: subsys mutex#24 ->&k->k_lock FD: 4 BD: 1 +.+.: subsys mutex#25 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 ....: netevent_notif_chain.lock FD: 50 BD: 1 +.+.: clients_rwsem ->fs_reclaim ->clients.xa_lock FD: 2 BD: 2 +.+.: clients.xa_lock ->pool_lock#2 FD: 28 BD: 8 .+.+: devices_rwsem ->&rq->__lock FD: 1 BD: 1 +.+.: (blocking_lsm_notifier_chain).rwsem FD: 111 BD: 45 ++++: (inetaddr_chain).rwsem ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->fib_info_lock ->&c->lock ->&dir->lock#2 ->&____s->seqcount ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&net->sctp.local_addr_lock ->&n->list_lock ->rlock-AF_NETLINK ->&rq->__lock ->rcu_node_0 ->remove_cache_srcu ->&ipvlan->addrs_lock ->&____s->seqcount#2 ->quarantine_lock ->&tbl->lock ->class ->(&tbl->proxy_timer) ->&base->lock ->krc.lock ->mmu_notifier_invalidate_range_start ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 8 ....: inet6addr_chain.lock FD: 1 BD: 1 +.+.: buses_mutex FD: 1 BD: 1 +.+.: offload_lock FD: 1 BD: 1 +...: inetsw_lock FD: 58 BD: 2 +.+.: pcpu_balance_work ->pcpu_alloc_mutex ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 47 +.+.: ptype_lock FD: 873 BD: 1 +.+.: (wq_completion)events_power_efficient ->(work_completion)(&(&tbl->managed_work)->work) ->(check_lifetime_work).work ->(work_completion)(&(&cache_cleaner)->work) ->(work_completion)(&(&ops->cursor_work)->work) ->(work_completion)(&(&hub->init_work)->work) ->(work_completion)(&(&gc_work->dwork)->work) ->(gc_work).work ->(work_completion)(&(&tbl->gc_work)->work) ->(crda_timeout).work ->&rq->__lock FD: 44 BD: 2 +.+.: (work_completion)(&(&tbl->managed_work)->work) ->&tbl->lock FD: 43 BD: 202 +.-.: &tbl->lock ->&obj_hash[i].lock ->&base->lock ->&n->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->pool_lock#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->&dir->lock#2 ->krc.lock ->&c->lock ->batched_entropy_u32.lock ->&____s->seqcount ->tk_core.seq.seqcount ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock ->init_task.mems_allowed_seq.seqcount ->&meta->lock FD: 28 BD: 2 +.+.: (check_lifetime_work).work ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 45 +.+.: &net->rules_mod_lock FD: 1 BD: 1 +.+.: tcp_ulp_list_lock FD: 1 BD: 1 +...: xfrm_state_afinfo_lock FD: 1 BD: 1 +.+.: xfrm_policy_afinfo_lock FD: 1 BD: 1 +...: xfrm_input_afinfo_lock FD: 15 BD: 653 ..-.: krc.lock ->&obj_hash[i].lock ->hrtimer_bases.lock ->&base->lock FD: 1 BD: 46 +...: k-slock-AF_INET/1 FD: 51 BD: 1 +.+.: (wq_completion)events_highpri ->(work_completion)(&(&krcp->page_cache_work)->work) FD: 50 BD: 2 +.+.: (work_completion)(&(&krcp->page_cache_work)->work) ->fs_reclaim ->&____s->seqcount ->krc.lock FD: 1 BD: 2 +.+.: &hashinfo->lock FD: 1 BD: 1 +.+.: tcp_cong_list_lock FD: 1 BD: 1 +.+.: mptcp_sched_list_lock FD: 2 BD: 7 +.+.: cache_list_lock ->&cd->hash_lock FD: 26 BD: 2 +.+.: (work_completion)(&(&cache_cleaner)->work) ->cache_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 1 BD: 1 +.+.: (rpc_pipefs_notifier_list).rwsem FD: 1 BD: 1 +.+.: svc_xprt_class_lock FD: 39 BD: 1 +.+.: xprt_list_lock ->(console_sem).lock FD: 129 BD: 3 ++++: umhelper_sem ->usermodehelper_disabled_waitq.lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->subsys mutex#73 ->fw_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&rq->__lock ->&cfs_rq->removed.lock ->&x->wait#22 ->&base->lock ->(&timer.timer) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start FD: 1 BD: 4 ....: usermodehelper_disabled_waitq.lock FD: 26 BD: 412 +.+.: &dentry->d_lock/1 ->&lru->node[i].lock FD: 84 BD: 3 .+.+: sb_writers#2 ->mount_lock ->&sb->s_type->i_mutex_key/1 ->&sb->s_type->i_mutex_key FD: 80 BD: 4 +.+.: &sb->s_type->i_mutex_key/1 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->tomoyo_ss ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&sb->s_type->i_mutex_key FD: 1 BD: 2 +.+.: tomoyo_log_lock FD: 1 BD: 2 ....: tomoyo_log_wait.lock FD: 54 BD: 110 +.+.: &wb->list_lock ->&sb->s_type->i_lock_key#2 ->&sb->s_type->i_lock_key#23 ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->&sb->s_type->i_lock_key#5 ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#3 ->&sb->s_type->i_lock_key#27 FD: 148 BD: 2 +.+.: (work_completion)(&sub_info->work) ->&sighand->siglock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->batched_entropy_u64.lock ->&obj_hash[i].lock ->init_files.file_lock ->init_fs.lock ->&p->alloc_lock ->lock ->pidmap_lock ->cgroup_threadgroup_rwsem ->input_pool.lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->&sig->wait_chldexit ->tasklist_lock ->&prev->lock ->&(&sig->stats_lock)->lock ->css_set_lock ->&x->wait#17 ->&____s->seqcount#2 ->remove_cache_srcu ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: umh_sysctl_lock FD: 1 BD: 3 +.+.: &drv->dynids.lock FD: 162 BD: 2 +.+.: &tsk->futex_exit_mutex ->&p->pi_lock ->&rq->__lock ->&mm->mmap_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 27 BD: 1 +.+.: &child->perf_event_mutex ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 93 ....: &pid->wait_pidfd FD: 25 BD: 98 ....: &sig->wait_chldexit ->&p->pi_lock FD: 28 BD: 98 ....: &(&sig->stats_lock)->lock ->&____s->seqcount#4 FD: 27 BD: 100 ....: &____s->seqcount#4 ->pidmap_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#26 FD: 141 BD: 1 +.+.: subsys mutex#27 ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->platform_devid_ida.xa_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#4 ->wakeup_ida.xa_lock ->gdp_mutex ->subsys mutex#13 ->events_lock ->rtcdev_lock FD: 1 BD: 1 +.+.: subsys mutex#28 FD: 1 BD: 90 +.+.: subsys mutex#29 FD: 1 BD: 4 +.+.: key_user_lock FD: 1 BD: 4 +.+.: key_serial_lock FD: 5 BD: 5 +.+.: key_construction_mutex ->&obj_hash[i].lock ->pool_lock#2 ->keyring_name_lock FD: 56 BD: 3 +.+.: &type->lock_class ->keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->key_user_lock ->crngs.lock ->key_serial_lock ->key_construction_mutex ->ima_keys_lock ->&c->lock ->&____s->seqcount FD: 52 BD: 4 +.+.: keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->root_key_user.lock ->key_construction_mutex FD: 1 BD: 1 ....: &pgdat->kswapd_wait FD: 1 BD: 1 +.+.: drivers_lock FD: 91 BD: 1 +.+.: damon_dbgfs_lock ->fs_reclaim ->pool_lock#2 ->damon_ops_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 75 BD: 1 +.+.: &type->s_umount_key#19/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#17 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#17 ->&dentry->d_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#20/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 8 +.+.: &sb->s_type->i_lock_key#18 ->&dentry->d_lock FD: 1 BD: 1 +.+.: configfs_subsystem_mutex FD: 71 BD: 1 +.+.: &sb->s_type->i_mutex_key#6/1 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]/2 ->&sb->s_type->i_mutex_key#7/2 ->&default_group_class[depth - 1]#3 FD: 1 BD: 9 +.+.: configfs_dirent_lock FD: 68 BD: 2 +.+.: &default_group_class[depth - 1]/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#4/2 FD: 49 BD: 1 +.+.: ecryptfs_daemon_hash_mux ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 ....: &ecryptfs_kthread_ctl.wait FD: 2 BD: 1 +.+.: ecryptfs_msg_ctx_lists_mux ->&ecryptfs_msg_ctx_arr[i].mux FD: 1 BD: 2 +.+.: &ecryptfs_msg_ctx_arr[i].mux FD: 38 BD: 1 +.+.: put_task_map-wait-type-override ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&base->lock ->task_group_lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->css_set_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 1 +.+.: nfs_version_lock FD: 77 BD: 1 ++++: key_types_sem ->(console_sem).lock ->asymmetric_key_parsers_sem ->&type->lock_class ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: pnfs_spinlock FD: 24 BD: 5 +.+.: &sn->pipefs_sb_lock ->&rq->__lock FD: 1 BD: 1 +.+.: nls_lock FD: 30 BD: 1 ..-.: &(&cache_cleaner)->timer FD: 36 BD: 2 +.+.: (work_completion)(&p->wq) ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->quarantine_lock ->&base->lock ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: jffs2_compressor_list_lock FD: 1 BD: 1 +.+.: next_tag_value_lock FD: 12 BD: 1 +.-.: (&tcp_orphan_timer) ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 1 ....: log_redrive_lock FD: 2 BD: 1 ....: &TxAnchor.LazyLock ->jfs_commit_thread_wait.lock FD: 1 BD: 2 ....: jfs_commit_thread_wait.lock FD: 1 BD: 1 +.+.: jfsTxnLock FD: 39 BD: 1 +.+.: ocfs2_stack_lock ->(console_sem).lock FD: 1 BD: 1 +.+.: o2hb_callback_sem FD: 1 BD: 1 +.+.: o2net_handler_lock FD: 3 BD: 1 +.+.: subsys mutex#30 ->&k->k_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#21/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&n->list_lock ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 37 BD: 5 +.+.: &sb->s_type->i_lock_key#19 ->&dentry->d_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#22/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#20 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&c->lock ->&____s->seqcount FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#20 ->&dentry->d_lock FD: 1 BD: 1 +.+.: cipso_v4_doi_list_lock FD: 224 BD: 51 +.+.: nf_hook_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&c->lock ->stock_lock ->&____s->seqcount#2 ->cpu_hotplug_lock ->&n->list_lock ->remove_cache_srcu ->&rq->__lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->nf_hook_mutex.wait_lock FD: 49 BD: 1 +.+.: alg_types_sem ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: dma_list_mutex FD: 68 BD: 2 ++++: asymmetric_key_parsers_sem ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->crypto_alg_sem ->&obj_hash[i].lock ->(crypto_chain).rwsem ->&x->wait#20 ->&base->lock ->&rq->__lock ->(&timer.timer) ->&c->lock ->quarantine_lock FD: 859 BD: 1 +.+.: blkcg_pol_register_mutex ->blkcg_pol_mutex ->cgroup_mutex FD: 1 BD: 2 +.+.: elv_list_lock FD: 53 BD: 1 +.+.: crc_t10dif_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 53 BD: 1 +.+.: crc64_rocksoft_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: ts_mod_lock FD: 1 BD: 1 +.+.: pci_ep_cfs_subsys.su_mutex FD: 64 BD: 1 +.+.: &default_group_class[depth - 1]#2/1 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#7/2 FD: 1 BD: 3 +.+.: &sb->s_type->i_mutex_key#7/2 FD: 1 BD: 1 +.+.: pci_epf_mutex FD: 230 BD: 1 +.+.: ipmi_interfaces_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->wq_pool_mutex ->&base->lock ->panic_notifier_list.lock FD: 2 BD: 1 +.+.: smi_watchers_mutex ->&ipmi_interfaces_srcu FD: 1 BD: 3 .+.?: &ipmi_interfaces_srcu FD: 1 BD: 1 +.+.: smi_infos_lock FD: 4 BD: 104 ....: mask_lock ->tmp_mask_lock FD: 3 BD: 105 ....: tmp_mask_lock ->tmpmask_lock ->&its->lock FD: 1 BD: 1 ....: &desc->wait_for_threads FD: 3 BD: 2 +.+.: subsys mutex#31 ->&k->k_lock FD: 26 BD: 3 +.+.: &dev->mutex#2 ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock FD: 107 BD: 1 +.+.: register_count_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&rq->__lock FD: 1 BD: 1 ....: thermal_cdev_ida.xa_lock FD: 1 BD: 1 ....: cpufreq_driver_lock FD: 3 BD: 1 +.+.: subsys mutex#32 ->&k->k_lock FD: 49 BD: 1 +.+.: scmi_requested_devices_mtx ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 ....: virtio_index_ida.xa_lock FD: 1 BD: 1 +.+.: subsys mutex#33 FD: 1 BD: 1 +.+.: vdpa_dev_lock FD: 75 BD: 3 ++++: &type->i_mutex_dir_key#2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&c->lock ->&____s->seqcount ->namespace_sem ->&sem->wait_lock ->&rq->__lock ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock FD: 3 BD: 1 +.+.: subsys mutex#34 ->&k->k_lock FD: 30 BD: 1 ..-.: lib/debugobjects.c:101 FD: 26 BD: 2 +.+.: (debug_obj_work).work ->pool_lock#2 ->&rq->__lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 6 +.+.: (work_completion)(&buf->work) FD: 1 BD: 1 ....: rng_index_ida.xa_lock FD: 121 BD: 4 +.+.: &md->mutex ->pci_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#6 ->&its->dev_alloc_lock ->&domain->mutex ->&irq_desc_lock_class ->tmpmask_lock ->&its->lock ->&root->kernfs_rwsem ->lock ->&obj_hash[i].lock ->sparse_irq_lock ->vmap_area_lock ->purge_vmap_area_lock ->&c->lock ->&____s->seqcount FD: 2 BD: 5 +.+.: &xa->xa_lock#6 ->pool_lock#2 FD: 51 BD: 8 +.+.: &its->dev_alloc_lock ->&its->lock ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->lpi_range_lock ->&c->lock FD: 1 BD: 106 ....: tmpmask_lock FD: 1 BD: 1 +.+.: &dev->vqs_list_lock FD: 1 BD: 1 ....: &vp_dev->lock FD: 52 BD: 1 +.+.: rng_mutex ->&x->wait#13 ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->&obj_hash[i].lock FD: 1 BD: 2 ....: &x->wait#13 FD: 25 BD: 2 -.-.: &x->wait#14 ->&p->pi_lock FD: 26 BD: 1 +.+.: reading_mutex ->&x->wait#14 ->&rq->__lock FD: 1 BD: 1 ....: &dev->config_lock FD: 2 BD: 1 +.-.: drivers/char/random.c:1010 ->input_pool.lock FD: 1 BD: 1 ....: &dev->managed.lock FD: 75 BD: 1 +.+.: &type->s_umount_key#23/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#21 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#21 ->&dentry->d_lock FD: 2 BD: 217 ....: drm_minor_lock ->pool_lock#2 FD: 3 BD: 3 +.+.: subsys mutex#35 ->&k->k_lock FD: 49 BD: 16 +.+.: &dev->mode_config.idr_mutex ->fs_reclaim ->pool_lock#2 FD: 91 BD: 12 +.+.: crtc_ww_class_acquire ->crtc_ww_class_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount FD: 90 BD: 13 +.+.: crtc_ww_class_mutex ->reservation_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.blob_lock ->&crtc->commit_lock ->reservation_ww_class_mutex ->tk_core.seq.seqcount ->&vkms_out->lock ->&dev->vbl_lock ->&x->wait#15 ->(work_completion)(&vkms_state->composer_work) ->&base->lock ->&rq->__lock ->(&timer.timer) ->(work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 14 +.+.: &dev->mode_config.blob_lock FD: 1 BD: 1 ....: &xa->xa_lock#7 FD: 1 BD: 1 ....: &xa->xa_lock#8 FD: 1 BD: 15 ....: &dev->mode_config.connector_list_lock FD: 17 BD: 17 ..-.: &dev->vbl_lock ->&dev->vblank_time_lock FD: 133 BD: 1 .+.+: drm_connector_list_iter ->&dev->mode_config.connector_list_lock ->fs_reclaim ->pool_lock#2 ->&connector->mutex FD: 131 BD: 2 +.+.: &connector->mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&rq->__lock ->&k->k_lock ->subsys mutex#35 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&dev->mode_config.idr_mutex ->connector_list_lock FD: 1 BD: 3 +.+.: connector_list_lock FD: 1 BD: 1 +.+.: &dev->filelist_mutex FD: 173 BD: 1 +.+.: &dev->clientlist_mutex ->&helper->lock ->registration_lock ->(console_sem).lock ->kernel_fb_helper_lock FD: 120 BD: 8 +.+.: &helper->lock ->fs_reclaim ->pool_lock#2 ->&client->modeset_mutex ->&obj_hash[i].lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&mgr->vm_lock ->&dev->object_name_lock ->&node->vm_lock ->&file_private->table_lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.fb_lock ->&file->fbs_lock ->&prime_fpriv->lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock ->&dev->master_mutex ->reservation_ww_class_mutex ->&lock->wait_lock ->&rq->__lock FD: 93 BD: 10 +.+.: &client->modeset_mutex ->&dev->mode_config.mutex ->fs_reclaim ->pool_lock#2 ->crtc_ww_class_acquire FD: 92 BD: 11 +.+.: &dev->mode_config.mutex ->crtc_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 2 BD: 9 +.+.: &mgr->vm_lock ->pool_lock#2 FD: 40 BD: 9 +.+.: &dev->object_name_lock ->lock FD: 10 BD: 217 +.+.: &file_private->table_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock FD: 4 BD: 9 +.+.: &node->vm_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 9 +.+.: &dev->mode_config.fb_lock FD: 1 BD: 9 +.+.: &file->fbs_lock FD: 1 BD: 9 +.+.: &prime_fpriv->lock FD: 171 BD: 2 +.+.: registration_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#11 ->vt_switch_mutex ->(console_sem).lock ->console_lock FD: 49 BD: 3 +.+.: vt_switch_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 6 +.+.: &fb_info->lock FD: 1 BD: 6 ....: vt_event_lock FD: 94 BD: 9 +.+.: &dev->master_mutex ->&client->modeset_mutex FD: 1 BD: 14 +.+.: &crtc->commit_lock FD: 23 BD: 307 -...: &xa->xa_lock#9 ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->&n->list_lock ->key#11 ->&s->s_inode_wblist_lock ->&base->lock ->key#12 ->&wb->work_lock ->key#13 ->stock_lock ->&xa->xa_lock#5 ->&pl->lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 94 +.+.: &info->lock FD: 14 BD: 718 +.+.: lock#4 ->&lruvec->lru_lock ->&obj_hash[i].lock ->&base->lock FD: 2 BD: 719 ....: &lruvec->lru_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 32 BD: 14 -.-.: &vkms_out->lock ->&dev->event_lock FD: 31 BD: 15 -.-.: &dev->event_lock ->&dev->vbl_lock ->&____s->seqcount#5 ->&x->wait#15 ->&obj_hash[i].lock ->pool_lock#2 ->&dev->vblank_time_lock ->&vblank->queue ->&base->lock FD: 1 BD: 20 ----: &____s->seqcount#5 FD: 25 BD: 16 -...: &x->wait#15 ->&p->pi_lock FD: 16 BD: 18 -.-.: &dev->vblank_time_lock ->tk_core.seq.seqcount ->&(&vblank->seqlock)->lock ->&obj_hash[i].lock ->hrtimer_bases.lock FD: 2 BD: 19 -.-.: &(&vblank->seqlock)->lock ->&____s->seqcount#5 FD: 1 BD: 14 +.+.: (work_completion)(&vkms_state->composer_work) FD: 1 BD: 10 ....: &helper->damage_lock FD: 122 BD: 2 +.+.: (work_completion)(&helper->damage_work) ->&helper->damage_lock ->&helper->lock FD: 1 BD: 16 -.-.: &vblank->queue FD: 1 BD: 14 +.+.: (work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 175 +.+.: &lock->wait_lock FD: 1 BD: 2 +.+.: kernel_fb_helper_lock FD: 1 BD: 1 +.+.: drivers_lock#2 FD: 1 BD: 1 +.+.: devices_lock FD: 1 BD: 8 ....: blk_queue_ida.xa_lock FD: 2 BD: 12 +.+.: &xa->xa_lock#10 ->pool_lock#2 FD: 16 BD: 261 ....: &q->queue_lock ->&blkcg->lock ->pool_lock#2 ->pcpu_lock ->&obj_hash[i].lock ->percpu_counters_lock ->&c->lock ->&____s->seqcount FD: 2 BD: 262 ....: &blkcg->lock ->pool_lock#2 FD: 1 BD: 14 +.+.: &bdev->bd_size_lock FD: 3 BD: 12 +.+.: subsys mutex#36 ->&k->k_lock FD: 246 BD: 8 +.+.: &q->sysfs_dir_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&q->sysfs_lock ->&obj_hash[i].lock FD: 245 BD: 9 +.+.: &q->sysfs_lock ->&q->debugfs_mutex ->&q->unused_hctx_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&obj_hash[i].lock ->cpu_hotplug_lock ->fs_reclaim ->&xa->xa_lock#11 ->&____s->seqcount ->pcpu_alloc_mutex ->&q->rq_qos_mutex ->&stats->lock ->&c->lock ->lock ->&root->kernfs_rwsem FD: 89 BD: 11 +.+.: &q->debugfs_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 5 BD: 23 ....: percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 3 BD: 8 +.+.: subsys mutex#37 ->&k->k_lock FD: 1 BD: 8 ....: cgwb_lock FD: 1 BD: 8 +...: bdi_lock FD: 52 BD: 263 +.+.: inode_hash_lock ->&sb->s_type->i_lock_key#3 ->&sb->s_type->i_lock_key#22 ->&s->s_inode_list_lock ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 FD: 2 BD: 10 +.+.: bdev_lock ->&bdev->bd_holder_lock FD: 138 BD: 9 +.+.: &disk->open_mutex ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&xa->xa_lock#9 ->lock#4 ->mmu_notifier_invalidate_range_start ->&c->lock ->&mapping->i_private_lock ->tk_core.seq.seqcount ->&ret->b_uptodate_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&rq->__lock ->&folio_wait_table[i] ->&cfs_rq->removed.lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&s->s_inode_list_lock ->pcpu_alloc_mutex ->&bdev->bd_size_lock ->&x->wait#9 ->ext_devt_ida.xa_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&k->k_lock ->subsys mutex#36 ->&xa->xa_lock#10 ->inode_hash_lock ->bdev_lock ->&lo->lo_mutex ->nbd_index_mutex ->&nbd->config_lock ->&new->lock ->&lock->wait_lock FD: 1 BD: 99 +.+.: &mapping->i_private_lock FD: 26 BD: 12 -...: &ret->b_uptodate_lock ->bit_wait_table + i FD: 1 BD: 714 +.+.: lock#5 FD: 49 BD: 1 +.+.: loop_ctl_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 10 +.+.: &q->unused_hctx_lock FD: 2 BD: 10 +.+.: &xa->xa_lock#11 ->pool_lock#2 FD: 1 BD: 7 +.+.: &set->tag_list_lock FD: 6 BD: 11 +.+.: &q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->mq_freeze_wq FD: 1 BD: 24 ....: &q->mq_freeze_wq FD: 94 BD: 10 +.+.: &q->rq_qos_mutex ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->debugfs_mutex ->set->srcu FD: 1 BD: 10 ....: &stats->lock FD: 60 BD: 10 +.+.: nbd_index_mutex ->fs_reclaim ->pool_lock#2 ->&nbd->config_lock FD: 1 BD: 17 .+.+: set->srcu FD: 29 BD: 3 +.+.: (work_completion)(&(&q->requeue_work)->work) ->&q->requeue_lock ->&hctx->lock ->&__ctx->lock FD: 16 BD: 3 +.+.: (work_completion)(&(&hctx->run_work)->work) FD: 30 BD: 1 ..-.: &(&ops->cursor_work)->timer FD: 32 BD: 2 +.+.: (work_completion)(&(&ops->cursor_work)->work) ->(console_sem).lock ->&obj_hash[i].lock ->&base->lock FD: 263 BD: 1 +.+.: zram_index_mutex ->fs_reclaim ->pool_lock#2 ->blk_queue_ida.xa_lock ->&obj_hash[i].lock ->pcpu_alloc_mutex ->bio_slab_lock ->&c->lock ->&____s->seqcount ->percpu_counters_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#10 ->lock ->&q->queue_lock ->&x->wait#9 ->&bdev->bd_size_lock ->&k->list_lock ->gdp_mutex ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->percpu_ref_switch_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#37 ->cgwb_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->bdi_lock ->inode_hash_lock ->(console_sem).lock FD: 3 BD: 1 +.+.: subsys mutex#38 ->&k->k_lock FD: 50 BD: 2 +.+.: &default_group_class[depth - 1]#3 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 2 BD: 1 +.+.: &lock ->nullb_indexes.xa_lock FD: 1 BD: 2 ....: nullb_indexes.xa_lock FD: 1 BD: 2 ....: nfc_index_ida.xa_lock FD: 145 BD: 4 +.+.: nfc_devlist_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->subsys mutex#39 ->&k->k_lock ->&genl_data->genl_data_mutex ->&____s->seqcount#2 ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->&x->wait#9 ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&rq->__lock ->&sem->wait_lock ->&n->list_lock ->&p->pi_lock ->quarantine_lock ->nfc_devlist_mutex.wait_lock ->dev_pm_qos_sysfs_mtx.wait_lock ->device_links_lock.wait_lock ->uevent_sock_mutex.wait_lock ->kn->active#4 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq FD: 26 BD: 5 +.+.: subsys mutex#39 ->&k->k_lock ->&rq->__lock FD: 1 BD: 2 +.+.: llcp_devices_lock FD: 1 BD: 53 ....: &rfkill->lock FD: 27 BD: 14 +.+.: subsys mutex#40 ->&k->k_lock ->&rq->__lock ->&lock->wait_lock FD: 121 BD: 6 +.+.: (work_completion)(&rfkill->sync_work) ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 14 +.+.: rfkill_global_mutex.wait_lock FD: 2 BD: 1 +.+.: dma_heap_minors.xa_lock ->pool_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#41 ->&k->k_lock FD: 1 BD: 1 +.+.: heap_list_lock FD: 4 BD: 1 +.+.: subsys mutex#42 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 +.+.: nvmf_hosts_mutex FD: 3 BD: 1 +.+.: subsys mutex#43 ->&k->k_lock FD: 1 BD: 1 +.+.: nvmf_transports_rwsem FD: 3 BD: 1 +.+.: subsys mutex#44 ->&k->k_lock FD: 1 BD: 3 ....: nvme_instance_ida.xa_lock FD: 3 BD: 3 +.+.: subsys mutex#45 ->&k->k_lock FD: 96 BD: 3 +.+.: pools_reg_lock ->pools_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem FD: 1 BD: 4 +.+.: pools_lock FD: 2 BD: 3 ....: &ctrl->lock ->&ctrl->state_wq FD: 1 BD: 4 ....: &ctrl->state_wq FD: 1 BD: 9 +.+.: &hctx->lock FD: 1 BD: 103 +.+.: &nvmeq->sq_lock FD: 25 BD: 6 ..-.: &x->wait#16 ->&p->pi_lock FD: 117 BD: 6 +.+.: nvme_subsystems_lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->subsys mutex#46 FD: 3 BD: 7 +.+.: subsys mutex#46 ->&k->k_lock FD: 1 BD: 7 +.+.: &xa->xa_lock#12 FD: 124 BD: 1 +.+.: &dev->shutdown_lock ->&md->mutex ->&desc->request_mutex ->&obj_hash[i].lock ->pool_lock#2 ->pci_lock ->fs_reclaim ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock FD: 1 BD: 104 ....: irq_resend_lock FD: 1 BD: 106 +.+.: &ent->pde_unload_lock FD: 1 BD: 6 ++++: &ctrl->namespaces_rwsem FD: 292 BD: 4 +.+.: (work_completion)(&ctrl->scan_work) ->&ctrl->scan_lock FD: 294 BD: 3 +.+.: (wq_completion)nvme-wq ->(work_completion)(&ctrl->scan_work) ->(work_completion)(&ctrl->async_event_work) ->(work_completion)(&barr->work) FD: 291 BD: 5 +.+.: &ctrl->scan_lock ->fs_reclaim ->pool_lock#2 ->&hctx->lock ->&x->wait#16 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->&c->lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&ctrl->namespaces_rwsem ->blk_queue_ida.xa_lock ->pcpu_alloc_mutex ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#10 ->lock ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->queue_lock ->&x->wait#9 ->nvme_subsystems_lock ->&subsys->lock ->&bdev->bd_size_lock ->ext_devt_ida.xa_lock ->&k->list_lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&k->k_lock ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->gdp_mutex ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#37 ->cgwb_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->bdi_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->nvme_ns_chr_minor_ida.xa_lock ->chrdevs_lock ->subsys mutex#50 ->&dentry->d_lock ->pool_lock ->quarantine_lock FD: 67 BD: 3 +.+.: &default_group_class[depth - 1]#4/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#5/2 FD: 1 BD: 1 +.+.: nvmet_config_sem FD: 3 BD: 1 +.+.: subsys mutex#47 ->&k->k_lock FD: 2 BD: 4 +.+.: (work_completion)(&ctrl->async_event_work) ->&nvmeq->sq_lock FD: 59 BD: 6 +.+.: &subsys->lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#12 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->pcpu_alloc_mutex FD: 1 BD: 10 ....: ext_devt_ida.xa_lock FD: 66 BD: 4 +.+.: &default_group_class[depth - 1]#5/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#6/2 FD: 65 BD: 5 +.+.: &default_group_class[depth - 1]#6/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#7 ->&default_group_class[depth - 1]#7/2 FD: 50 BD: 6 +.+.: &default_group_class[depth - 1]#7 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 1 BD: 6 +.+.: &default_group_class[depth - 1]#7/2 FD: 1 BD: 1 +.+.: backend_mutex FD: 1 BD: 1 +.+.: scsi_mib_index_lock FD: 1 BD: 1 +.+.: hba_lock FD: 49 BD: 1 +.+.: device_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: &hba->device_lock FD: 25 BD: 701 -.-.: &folio_wait_table[i] ->&p->pi_lock FD: 1 BD: 1 +.+.: part_parser_lock FD: 279 BD: 1 +.+.: mtd_table_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#48 ->devtree_lock ->nvmem_ida.xa_lock ->nvmem_cell_mutex ->&k->k_lock ->subsys mutex#49 ->nvmem_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->(console_sem).lock ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->mmu_notifier_invalidate_range_start ->blk_queue_ida.xa_lock ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#10 ->&q->mq_freeze_lock ->set->srcu ->percpu_ref_switch_lock ->&q->queue_lock ->&bdev->bd_size_lock ->elv_list_lock ->(work_completion)(&(&q->requeue_work)->work) ->(work_completion)(&(&hctx->run_work)->work) ->&q->debugfs_mutex ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->subsys mutex#37 ->cgwb_lock ->bdi_lock ->inode_hash_lock FD: 1 BD: 1 +.+.: chip_drvs_lock FD: 1 BD: 6 ....: (kmod_concurrent_max).lock FD: 25 BD: 8 ....: &x->wait#17 ->&p->pi_lock FD: 1 BD: 3 ....: &prev->lock FD: 3 BD: 2 +.+.: subsys mutex#48 ->&k->k_lock FD: 1 BD: 2 ....: nvmem_ida.xa_lock FD: 1 BD: 2 +.+.: nvmem_cell_mutex FD: 1 BD: 2 +.+.: subsys mutex#49 FD: 1 BD: 2 +.+.: nvmem_mutex FD: 1 BD: 45 +.+.: &bond->stats_lock FD: 33 BD: 59 ....: lweventlist_lock ->pool_lock#2 ->&dir->lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&base->lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 33 BD: 1 +.+.: (wq_completion)gid-cache-wq ->(work_completion)(&ndev_work->work) ->(work_completion)(&work->work) ->&rq->__lock FD: 31 BD: 2 +.+.: (work_completion)(&ndev_work->work) ->devices_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock FD: 850 BD: 2 +.+.: (linkwatch_work).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 73 +.+.: rtnl_mutex.wait_lock FD: 3 BD: 46 ..-.: once_lock ->crngs.lock FD: 229 BD: 2 +.+.: (work_completion)(&w->work) ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 45 ++++: (inet6addr_validator_chain).rwsem ->&rq->__lock FD: 24 BD: 45 ++++: (inetaddr_validator_chain).rwsem ->&rq->__lock FD: 1 BD: 6 ....: nvme_ns_chr_minor_ida.xa_lock FD: 3 BD: 6 +.+.: subsys mutex#50 ->&k->k_lock FD: 3 BD: 1 +.+.: subsys mutex#51 ->&k->k_lock FD: 1 BD: 1 +.+.: gpio_lookup_lock FD: 1 BD: 1 +.+.: mdio_board_lock FD: 1 BD: 1 +.+.: mode_list_lock FD: 1 BD: 1 +.+.: l3mdev_lock FD: 1 BD: 92 -.-.: &retval->lock FD: 44 BD: 1 +.+.: (wq_completion)gve ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->(work_completion)(&priv->service_task) FD: 43 BD: 2 +.+.: (work_completion)(&priv->service_task) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->(console_sem).lock ->lweventlist_lock ->&base->lock ->&dir->lock#2 FD: 1 BD: 1 +.+.: hnae3_common_lock FD: 3 BD: 1 +.+.: subsys mutex#52 ->&k->k_lock FD: 2 BD: 1 +.+.: compressor_list_lock ->pool_lock#2 FD: 1 BD: 5 ....: hwsim_netgroup_ida.xa_lock FD: 34 BD: 74 +.-.: hwsim_radio_lock ->pool_lock#2 ->&list->lock#16 ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->init_task.mems_allowed_seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 3 BD: 8 +.+.: subsys mutex#53 ->&k->k_lock FD: 51 BD: 46 +.+.: param_lock ->rate_ctrl_mutex ->disk_events_mutex ->&rq->__lock FD: 399 BD: 48 +.+.: &rdev->wiphy.mtx ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#54 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->nl_table_lock ->nl_table_wait.lock ->reg_requests_lock ->stack_depot_init_mutex ->pcpu_alloc_mutex ->&xa->xa_lock#4 ->net_rwsem ->&x->wait#9 ->subsys mutex#20 ->&dir->lock#2 ->&rq->__lock ->&cfs_rq->removed.lock ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->&fq->lock ->&local->iflist_mtx ->rlock-AF_NETLINK ->&rdev->bss_lock ->lweventlist_lock ->&base->lock ->&data->mutex ->&rdev->wiphy_work_lock ->&local->filter_lock ->&tn->lock ->failover_lock ->&n->list_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&idev->mc_lock ->&pnettable->lock ->smc_ib_devices.mutex ->&ndev->lock ->&wdev->event_lock ->&rdev->mgmt_registrations_lock ->(&dwork->timer) ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->mount_lock ->(&dwork->timer)#2 ->(work_completion)(&(&link->color_collision_detect_work)->work) ->rtnl_mutex.wait_lock ->&p->pi_lock ->key ->pcpu_lock ->percpu_counters_lock ->&list->lock#15 ->&ifibss->incomplete_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->tk_core.seq.seqcount ->hrtimer_bases.lock ->&list->lock#2 ->&sta->lock ->lock#6 ->&____s->seqcount#2 ->pool_lock ->&lock->wait_lock ->&sem->wait_lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&wdev->pmsr_lock ->krc.lock ->&local->active_txq_lock[i] ->(work_completion)(&sta->drv_deliver_wk) ->&local->queue_stop_reason_lock ->&x->wait#2 ->(&ifibss->timer) ->_xmit_ETHER ->(&local->dynamic_ps_timer) ->(&dwork->timer)#3 ->&list->lock#16 ->&wq->mutex ->cpu_hotplug_lock ->&rnp->exp_wq[2] ->bpf_devs_lock ->&in_dev->mc_tomb_lock ->class ->(&tbl->proxy_timer) ->&ul->lock ->&net->xdp.lock ->mirred_list_lock ->&nft_net->commit_mutex ->&ul->lock#2 ->&ent->pde_unload_lock ->&net->ipv6.addrconf_hash_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&idev->mc_report_lock ->&pnn->pndevs.lock ->&pnn->routes.lock ->target_list_lock ->kernfs_idr_lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->(&dwork->timer)#4 ->dev_pm_qos_sysfs_mtx.wait_lock ->mmu_notifier_invalidate_range_start ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->stock_lock ->rcu_state.exp_mutex.wait_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->(wq_completion)phy40 ->(wq_completion)phy39 ->(wq_completion)phy32 ->(wq_completion)phy31 ->(wq_completion)phy44 ->(wq_completion)phy43 ->(wq_completion)phy46 ->(wq_completion)phy45 ->(wq_completion)phy48 ->(wq_completion)phy47 ->(wq_completion)phy50 ->(wq_completion)phy49 ->(wq_completion)phy34 ->(wq_completion)phy33 ->(wq_completion)phy54 ->(wq_completion)phy53 ->(wq_completion)phy52 ->(wq_completion)phy51 ->(wq_completion)phy42 ->(wq_completion)phy41 ->(wq_completion)phy56 ->(wq_completion)phy55 ->(wq_completion)phy62 ->(wq_completion)phy61 ->(wq_completion)phy60 ->(wq_completion)phy59 ->(wq_completion)phy38 ->(wq_completion)phy37 ->(wq_completion)phy64 ->(wq_completion)phy63 ->(wq_completion)phy66 ->(wq_completion)phy65 ->(wq_completion)phy58 ->(wq_completion)phy57 ->(wq_completion)phy76 ->(wq_completion)phy75 ->uevent_sock_mutex.wait_lock FD: 3 BD: 49 +.+.: subsys mutex#54 ->&k->k_lock FD: 1 BD: 49 +.+.: reg_requests_lock FD: 7 BD: 73 +.-.: &fq->lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&zone->lock ->pool_lock#2 FD: 1 BD: 49 +.+.: &local->iflist_mtx FD: 3 BD: 45 +.+.: subsys mutex#55 ->&k->k_lock FD: 2 BD: 46 +.+.: &sdata->sec_mtx ->&sec->lock FD: 1 BD: 47 +...: &sec->lock FD: 1 BD: 45 +.+.: &local->iflist_mtx#2 FD: 49 BD: 1 +.+.: hwsim_phys_lock ->fs_reclaim ->pool_lock#2 FD: 49 BD: 1 +.+.: xdomain_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: ioctl_mutex FD: 1 BD: 1 +.+.: address_handler_list_lock FD: 1 BD: 1 +.+.: card_mutex FD: 3 BD: 1 +.+.: subsys mutex#56 ->&k->k_lock FD: 25 BD: 1 ....: &x->wait#18 ->&p->pi_lock FD: 27 BD: 2 ..-.: &txlock ->&list->lock#3 ->&txwq FD: 1 BD: 3 ..-.: &list->lock#3 FD: 25 BD: 3 ..-.: &txwq ->&p->pi_lock FD: 2 BD: 1 ....: &iocq[i].lock ->&ktiowq[i] FD: 1 BD: 2 ....: &ktiowq[i] FD: 1 BD: 1 ....: rcu_read_lock_bh FD: 1 BD: 57 +.-.: noop_qdisc.q.lock FD: 3 BD: 5 +.+.: subsys mutex#57 ->&k->k_lock FD: 178 BD: 1 +.+.: usb_bus_idr_lock ->(usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->hcd_root_hub_lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&dev->power.lock ->device_links_srcu ->&____s->seqcount ->&c->lock ->(console_sem).lock ->input_pool.lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#58 ->&x->wait#9 ->&vhci_hcd->vhci->lock ->&lock->wait_lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->(&timer.timer) ->quarantine_lock FD: 116 BD: 1 +.+.: table_lock ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->(console_sem).lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 2 BD: 1 +.-.: (&ipmi_timer) ->&ipmi_interfaces_srcu FD: 1 BD: 5 +.+.: mon_lock FD: 113 BD: 4 +.+.: usb_port_peer_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->dev_pm_qos_mtx ->component_mutex ->device_links_srcu ->dev_pm_qos_sysfs_mtx ->&rq->__lock ->sysfs_symlink_target_lock ->device_state_lock FD: 31 BD: 22 ....: device_state_lock ->kernfs_notify_lock FD: 37 BD: 20 ..-.: hcd_root_hub_lock ->hcd_urb_list_lock ->&bh->lock ->&p->pi_lock ->&dev->power.lock ->device_state_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 21 ..-.: hcd_urb_list_lock FD: 1 BD: 21 ..-.: &bh->lock FD: 14 BD: 100 ..-.: lock#6 ->kcov_remote_lock ->&kcov->lock FD: 12 BD: 135 ..-.: kcov_remote_lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&base->lock FD: 25 BD: 16 ..-.: &x->wait#19 ->&p->pi_lock FD: 1 BD: 4 +.+.: set_config_lock FD: 66 BD: 5 +.+.: hcd->bandwidth_mutex ->devtree_lock ->&obj_hash[i].lock ->&x->wait#9 ->&dev->power.lock ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->hcd_root_hub_lock ->&rq->__lock ->&x->wait#19 ->&____s->seqcount ->&c->lock ->&base->lock ->(&timer.timer) ->&dum_hcd->dum->lock ->&n->list_lock FD: 1 BD: 4 +.+.: &new_driver->dynids.lock FD: 38 BD: 18 ..-.: &dum_hcd->dum->lock ->hcd_root_hub_lock ->hcd_urb_list_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 FD: 67 BD: 11 +.+.: &hub->status_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->hcd_root_hub_lock ->fs_reclaim ->&dum_hcd->dum->lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&c->lock ->&____s->seqcount ->&vhci_hcd->vhci->lock ->&base->lock ->(&timer.timer) ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->quarantine_lock ->&n->list_lock ->&queue->lock ->remove_cache_srcu ->&____s->seqcount#2 ->rcu_node_0 FD: 1 BD: 5 +.+.: component_mutex FD: 1 BD: 4 +.+.: subsys mutex#58 FD: 69 BD: 2 +.+.: (work_completion)(&(&hub->init_work)->work) ->&rq->__lock ->&lock->wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 210 BD: 1 +.+.: (wq_completion)usb_hub_wq ->(work_completion)(&hub->events) FD: 209 BD: 2 +.+.: (work_completion)(&hub->events) ->lock#6 ->&rq->__lock ->&dev->power.lock FD: 38 BD: 3 ..-.: &hub->irq_urb_lock ->hcd_root_hub_lock FD: 1 BD: 3 ....: (&hub->irq_urb_retry) FD: 1 BD: 5 ....: hcd_urb_unlink_lock FD: 25 BD: 3 ..-.: usb_kill_urb_queue.lock ->&p->pi_lock FD: 44 BD: 17 +.+.: udc_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->udc_lock.wait_lock ->&queue->lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 3 +.+.: (work_completion)(&hub->tt.clear_work) FD: 3 BD: 1 +.+.: subsys mutex#59 ->&k->k_lock FD: 1 BD: 1 ....: gadget_id_numbers.xa_lock FD: 86 BD: 2 +.+.: (work_completion)(&gadget->work) ->&root->kernfs_rwsem ->kernfs_notify_lock ->&sem->wait_lock ->&p->pi_lock FD: 30 BD: 126 ..-.: kernfs_notify_lock FD: 59 BD: 2 +.+.: kernfs_notify_work ->kernfs_notify_lock ->&root->kernfs_supers_rwsem FD: 53 BD: 7 ++++: &root->kernfs_supers_rwsem ->inode_hash_lock ->&rq->__lock FD: 1 BD: 1 +.+.: subsys mutex#60 FD: 1 BD: 1 +.+.: func_lock FD: 1 BD: 1 +.+.: g_tf_lock FD: 1 BD: 13 ....: &vhci_hcd->vhci->lock FD: 2 BD: 2 ....: input_ida.xa_lock ->pool_lock#2 FD: 1 BD: 1 ....: &mousedev->mutex/1 FD: 30 BD: 4 ....: serio_event_lock ->pool_lock#2 FD: 56 BD: 1 +.+.: (wq_completion)events_long ->serio_event_work ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&br->gc_work)->work) ->(work_completion)(&br->mcast_gc_work) FD: 35 BD: 2 +.+.: serio_event_work ->serio_mutex FD: 34 BD: 3 +.+.: serio_mutex ->serio_event_lock ->&k->list_lock ->&k->k_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 ....: rtc_ida.xa_lock FD: 32 BD: 1 +.+.: &rtc->ops_lock ->(efi_runtime_lock).lock ->&obj_hash[i].lock ->&x->wait#12 ->&rq->__lock FD: 1 BD: 2 ....: platform_devid_ida.xa_lock FD: 1 BD: 2 ....: rtcdev_lock FD: 49 BD: 1 +.+.: g_smscore_deviceslock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: cx231xx_devlist_mutex FD: 1 BD: 1 +.+.: em28xx_devlist_mutex FD: 1 BD: 1 ....: pvr2_context_sync_data.lock FD: 1 BD: 10 +.+.: i2c_dev_list_lock FD: 3 BD: 10 +.+.: subsys mutex#61 ->&k->k_lock FD: 1 BD: 1 +.+.: subsys mutex#62 FD: 129 BD: 2 +.+.: dvbdev_register_lock ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->minor_rwsem ->&xa->xa_lock#13 ->&mdev->graph_mutex ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#63 FD: 130 BD: 1 +.+.: frontend_mutex ->fs_reclaim ->pool_lock#2 ->(console_sem).lock ->dvbdev_register_lock FD: 1 BD: 3 +.+.: minor_rwsem FD: 8 BD: 3 ....: &xa->xa_lock#13 ->&c->lock ->&____s->seqcount ->pool_lock#2 FD: 49 BD: 4 +.+.: &mdev->graph_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&rq->__lock FD: 3 BD: 3 +.+.: subsys mutex#63 ->&k->k_lock FD: 1 BD: 1 ....: &dmxdev->lock FD: 1 BD: 1 +.+.: &dvbdemux->mutex FD: 1 BD: 1 +.+.: media_devnode_lock FD: 1 BD: 1 +.+.: subsys mutex#64 FD: 24 BD: 1 +.+.: videodev_lock ->&rq->__lock FD: 3 BD: 1 +.+.: subsys mutex#65 ->&k->k_lock FD: 1 BD: 1 +.+.: vimc_sensor:396:(&vsensor->hdl)->_lock FD: 1 BD: 1 +.+.: &v4l2_dev->lock FD: 1 BD: 1 +.+.: vimc_debayer:581:(&vdebayer->hdl)->_lock FD: 1 BD: 1 +.+.: vimc_lens:61:(&vlens->hdl)->_lock FD: 59 BD: 1 +.+.: vivid_ctrls:1606:(hdl_user_gen)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock ->&zone->lock ->&obj_hash[i].lock FD: 50 BD: 1 +.+.: vivid_ctrls:1608:(hdl_user_vid)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 FD: 53 BD: 1 +.+.: vivid_ctrls:1610:(hdl_user_aud)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock ->&____s->seqcount FD: 57 BD: 1 +.+.: vivid_ctrls:1612:(hdl_streaming)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock FD: 51 BD: 1 +.+.: vivid_ctrls:1614:(hdl_sdtv_cap)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 51 BD: 1 +.+.: vivid_ctrls:1616:(hdl_loop_cap)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 1 BD: 1 +.+.: vivid_ctrls:1618:(hdl_fb)->_lock FD: 1 BD: 7 +.+.: vivid_ctrls:1620:(hdl_vid_cap)->_lock FD: 1 BD: 4 +.+.: vivid_ctrls:1622:(hdl_vid_out)->_lock FD: 1 BD: 5 +.+.: vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1627:(hdl_vbi_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1630:(hdl_radio_rx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1632:(hdl_radio_tx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1634:(hdl_sdr_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1636:(hdl_meta_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1638:(hdl_meta_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1640:(hdl_tch_cap)->_lock FD: 1 BD: 1 ....: &adap->kthread_waitq FD: 1 BD: 1 +.+.: &dev->cec_xfers_slock FD: 1 BD: 1 ....: &dev->kthread_waitq_cec FD: 1 BD: 1 +.+.: cec_devnode_lock FD: 1 BD: 1 +.+.: subsys mutex#66 FD: 5 BD: 1 +.+.: &adap->lock ->tk_core.seq.seqcount ->&adap->devnode.lock_fhs FD: 1 BD: 2 +.+.: &adap->devnode.lock_fhs FD: 1 BD: 1 ....: ptp_clocks_map.xa_lock FD: 3 BD: 1 +.+.: subsys mutex#67 ->&k->k_lock FD: 1 BD: 1 +.+.: pers_lock FD: 1 BD: 1 +.+.: _lock FD: 1 BD: 3 +.+.: dm_bufio_clients_lock FD: 1 BD: 1 +.+.: _ps_lock FD: 1 BD: 1 +.+.: _lock#2 FD: 1 BD: 1 +.+.: _lock#3 FD: 1 BD: 1 +.+.: register_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#68 ->&k->k_lock FD: 1 BD: 1 .+.+: bp_lock FD: 3 BD: 1 +.+.: subsys mutex#69 ->&k->k_lock FD: 14 BD: 1 +.-.: (&dsp_spl_tl) ->dsp_lock FD: 13 BD: 2 ..-.: dsp_lock ->iclock_lock ->&obj_hash[i].lock ->&base->lock FD: 4 BD: 3 ...-: iclock_lock ->tk_core.seq.seqcount FD: 50 BD: 45 +.+.: lock#7 ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#15 FD: 1 BD: 1 ....: iscsi_transport_lock FD: 3 BD: 1 +.+.: subsys mutex#70 ->&k->k_lock FD: 1 BD: 1 ....: &tx_task->waiting FD: 1 BD: 1 +.+.: link_ops_rwsem FD: 127 BD: 1 +.+.: disable_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#4 FD: 2 BD: 1 +.+.: protocol_lock ->pool_lock#2 FD: 72 BD: 1 +.+.: psinfo_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->(console_sem).lock ->pstore_sb_lock ->dump_list_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 58 BD: 3 +.+.: pstore_sb_lock ->&sb->s_type->i_mutex_key#12 FD: 1 BD: 2 ....: dump_list_lock FD: 1 BD: 1 +.+.: vsock_register_mutex FD: 1 BD: 1 +.+.: comedi_drivers_list_lock FD: 102 BD: 1 +.+.: cscfg_mutex ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&dev->power.lock ->dpm_list_mtx FD: 1 BD: 706 +.+.: icc_bw_lock FD: 3 BD: 9 +.+.: subsys mutex#71 ->&k->k_lock FD: 113 BD: 5 ++++: snd_ctl_layer_rwsem ->snd_ctl_led_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->sysfs_symlink_target_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->&n->list_lock ->&____s->seqcount#2 ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 6 +.+.: snd_card_mutex FD: 1 BD: 1 +.+.: snd_ioctl_rwsem FD: 49 BD: 5 +.+.: strings ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&c->lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: register_mutex FD: 127 BD: 6 +.+.: sound_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->&obj_hash[i].lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#71 ->&k->k_lock ->&cfs_rq->removed.lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->mmu_notifier_invalidate_range_start ->&x->wait#9 ->deferred_probe_mutex ->device_links_lock ->&sem->wait_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->uevent_sock_mutex.wait_lock FD: 136 BD: 1 +.+.: register_mutex#2 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->register_mutex ->sound_oss_mutex ->strings ->&entry->access ->info_mutex FD: 129 BD: 1 +.+.: register_mutex#3 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->clients_lock ->&c->lock ->&____s->seqcount FD: 1 BD: 5 ....: clients_lock FD: 2 BD: 1 +.+.: &client->ports_mutex ->&client->ports_lock FD: 1 BD: 5 .+.+: &client->ports_lock FD: 122 BD: 1 +.+.: register_mutex#4 ->fs_reclaim ->pool_lock#2 ->sound_oss_mutex FD: 121 BD: 6 +.+.: sound_oss_mutex ->fs_reclaim ->pool_lock#2 ->sound_loader_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#71 ->&cfs_rq->removed.lock ->&k->k_lock ->&n->list_lock ->&____s->seqcount#2 ->uevent_sock_mutex.wait_lock FD: 1 BD: 7 +.+.: sound_loader_lock FD: 52 BD: 1 .+.+: &grp->list_mutex/1 ->clients_lock ->&client->ports_lock ->register_lock#3 ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 2 BD: 1 +.+.: &grp->list_mutex#2 ->&grp->list_lock FD: 1 BD: 2 ....: &grp->list_lock FD: 63 BD: 2 +.+.: async_lookup_work ->fs_reclaim ->pool_lock#2 ->clients_lock ->&client->ports_lock ->snd_card_mutex ->(kmod_concurrent_max).lock ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->running_helpers_waitq.lock ->autoload_work ->&x->wait#10 FD: 1 BD: 2 ....: register_lock#3 FD: 116 BD: 4 ++++: &card->controls_rwsem ->&xa->xa_lock#14 ->fs_reclaim ->&card->ctl_files_rwlock ->snd_ctl_layer_rwsem ->pool_lock#2 ->&rq->__lock FD: 9 BD: 5 +.+.: &xa->xa_lock#14 ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 1 BD: 5 ....: &card->ctl_files_rwlock FD: 4 BD: 3 +.+.: autoload_work ->&k->list_lock ->&k->k_lock FD: 1 BD: 6 +.+.: snd_ctl_led_mutex FD: 1 BD: 1 +.+.: register_mutex#5 FD: 50 BD: 1 +.+.: client_mutex ->fs_reclaim ->pool_lock#2 ->&dev->devres_lock ->&c->lock ->&____s->seqcount FD: 1 BD: 49 +.+.: failover_lock FD: 8 BD: 2 +...: llc_sap_list_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 49 BD: 1 +.+.: act_id_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount FD: 1 BD: 1 +.+.: act_mod_lock FD: 1 BD: 1 +.+.: ife_mod_lock FD: 1 BD: 1 +.+.: cls_mod_lock FD: 1 BD: 1 +.+.: ematch_mod_lock FD: 1 BD: 1 +.+.: sock_diag_table_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_acct FD: 1 BD: 1 +.+.: nfnl_subsys_queue FD: 1 BD: 1 +.+.: nfnl_subsys_ulog FD: 1 BD: 5 +.+.: nf_log_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_osf FD: 32 BD: 1 +.+.: nf_sockopt_mutex ->&rq->__lock ->nf_sockopt_mutex.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: nfnl_subsys_ctnetlink FD: 1 BD: 1 +.+.: nfnl_subsys_ctnetlink_exp FD: 1 BD: 5 +.+.: nf_ct_ecache_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_cttimeout FD: 1 BD: 1 +.+.: nfnl_subsys_cthelper FD: 1 BD: 1 +.+.: nf_ct_helper_mutex FD: 1 BD: 1 +...: nf_conntrack_expect_lock FD: 1 BD: 1 +.+.: nf_ct_nat_helpers_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_nftables FD: 1 BD: 1 +.+.: nfnl_subsys_nftcompat FD: 1420 BD: 1 +.+.: masq_mutex ->pernet_ops_rwsem ->(inetaddr_chain).rwsem ->inet6addr_chain.lock FD: 164 BD: 5 +.+.: &xt[i].mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&mm->mmap_lock ->free_vmap_area_lock ->vmap_area_lock ->&per_cpu(xt_recseq, i) ->&obj_hash[i].lock ->purge_vmap_area_lock ->init_mm.page_table_lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->remove_cache_srcu ->rcu_node_0 ->&lock->wait_lock ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->stock_lock ->pool_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock FD: 27 BD: 79 +.+.: &tn->lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 3 BD: 1 +.+.: subsys mutex#72 ->&k->k_lock FD: 24 BD: 5 +.+.: nfnl_subsys_ipset ->&rq->__lock FD: 1 BD: 1 +.+.: ip_set_type_mutex FD: 58 BD: 5 +.+.: ipvs->est_mutex ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->pcpu_lock ->&obj_hash[i].lock ->&n->list_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock FD: 1 BD: 1 +.+.: ip_vs_sched_mutex FD: 49 BD: 5 +.+.: __ip_vs_app_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 1 +.+.: ip_vs_pe_mutex FD: 1 BD: 1 +.+.: tunnel4_mutex FD: 1 BD: 1 +.+.: xfrm4_protocol_mutex FD: 1 BD: 1 +.+.: inet_diag_table_mutex FD: 1 BD: 1 +...: xfrm_km_lock FD: 1 BD: 1 +.+.: xfrm6_protocol_mutex FD: 1 BD: 1 +.+.: tunnel6_mutex FD: 1 BD: 1 +.+.: xfrm_if_cb_lock FD: 1 BD: 1 +...: inetsw6_lock FD: 1 BD: 6 +.+.: &hashinfo->lock#2 FD: 16 BD: 5 +.+.: &net->ipv6.ip6addrlbl_table.lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 132 BD: 52 +.+.: &idev->mc_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&dev_addr_list_lock_key ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->batched_entropy_u32.lock ->&base->lock ->&n->list_lock ->krc.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&bridge_netdev_addr_lock_key ->&dev_addr_list_lock_key#2 ->&batadv_netdev_addr_lock_key ->&rq->__lock ->remove_cache_srcu ->&vlan_netdev_addr_lock_key ->&macvlan_netdev_addr_lock_key ->&dev_addr_list_lock_key#3 ->&bridge_netdev_addr_lock_key/1 ->&dev_addr_list_lock_key/1 ->pcpu_lock ->&dev_addr_list_lock_key#2/1 ->rcu_node_0 ->_xmit_ETHER/1 ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&dev_addr_list_lock_key#3/1 ->&macsec_netdev_addr_lock_key/1 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->key ->percpu_counters_lock FD: 8 BD: 53 +...: &dev_addr_list_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 42 BD: 65 +...: _xmit_ETHER ->&local->filter_lock ->&rdev->wiphy_work_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->(console_sem).lock ->console_owner_lock ->console_owner ->&____s->seqcount#2 ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 852 BD: 1 +.+.: (wq_completion)ipv6_addrconf ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->(work_completion)(&(&ifa->dad_work)->work) FD: 850 BD: 6 +.+.: (work_completion)(&(&net->ipv6.addr_chk_work)->work) ->rtnl_mutex FD: 25 BD: 3 ....: &x->wait#20 ->&p->pi_lock FD: 47 BD: 79 ++--: &ndev->lock ->&ifa->lock ->pool_lock#2 ->&dir->lock#2 ->pcpu_lock ->&tb->tb6_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u32.lock ->&base->lock FD: 9 BD: 1 +.+.: stp_proto_mutex ->llc_sap_list_lock FD: 1 BD: 1 ....: switchdev_notif_chain.lock FD: 1 BD: 45 ++++: (switchdev_blocking_notif_chain).rwsem FD: 1 BD: 1 +.+.: br_ioctl_mutex FD: 230 BD: 8 +.+.: nf_ct_proto_mutex ->defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->defrag6_mutex ->&rq->__lock ->nf_ct_proto_mutex.wait_lock FD: 163 BD: 5 +.+.: ebt_mutex ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&____s->seqcount ->stock_lock ->&c->lock ->&rq->__lock ->ebt_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&n->list_lock ->rcu_node_0 ->&____s->seqcount#2 FD: 1 BD: 1 +.+.: dsa_tag_drivers_lock FD: 1 BD: 1 +...: protocol_list_lock FD: 1 BD: 1 +...: linkfail_lock FD: 1 BD: 1 +...: rose_neigh_list_lock FD: 1 BD: 1 +.+.: proto_tab_lock#2 FD: 1 BD: 53 ++++: chan_list_lock FD: 1 BD: 2 +.+.: l2cap_sk_list.lock FD: 3 BD: 1 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->chan_list_lock FD: 1 BD: 2 +...: slock-AF_BLUETOOTH-BTPROTO_L2CAP FD: 1 BD: 1 ....: rfcomm_wq.lock FD: 1 BD: 1 +.+.: rfcomm_mutex FD: 1 BD: 1 +.+.: auth_domain_lock FD: 1 BD: 1 +.+.: registered_mechs_lock FD: 1 BD: 1 ....: atm_dev_notify_chain.lock FD: 1 BD: 1 +.+.: proto_tab_lock#3 FD: 850 BD: 1 +.+.: vlan_ioctl_mutex ->&mm->mmap_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 1 +.+.: rds_info_lock FD: 39 BD: 1 +.+.: rds_trans_sem ->(console_sem).lock FD: 1 BD: 1 ....: &id_priv->lock FD: 2 BD: 46 +.+.: &xa->xa_lock#15 ->pool_lock#2 FD: 48 BD: 48 +.+.: k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&table->hash[i].lock ->k-clock-AF_INET6 ->&icsk->icsk_accept_queue.rskq_lock#2 ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 FD: 25 BD: 50 +...: k-slock-AF_INET6 ->pool_lock#2 ->&obj_hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->key#23 FD: 1 BD: 69 ++..: k-clock-AF_INET6 FD: 18 BD: 64 +.-.: &tcp_hashinfo.bhash[i].lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&tcp_hashinfo.bhash2[i].lock ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 FD: 17 BD: 65 +.-.: &tcp_hashinfo.bhash2[i].lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->&obj_hash[i].lock ->batched_entropy_u8.lock ->&hashinfo->ehash_locks[i] ->stock_lock ->&____s->seqcount#2 FD: 1 BD: 51 +.+.: &h->lhash2[i].lock FD: 1 BD: 5 +...: &list->lock#4 FD: 1 BD: 6 +...: k-clock-AF_TIPC FD: 33 BD: 5 +.+.: k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&tn->nametbl_lock ->&obj_hash[i].lock ->k-clock-AF_TIPC ->&rq->__lock FD: 1 BD: 6 +...: k-slock-AF_TIPC FD: 18 BD: 6 +...: &tn->nametbl_lock ->pool_lock#2 ->&service->lock ->&c->lock ->&____s->seqcount ->&nt->cluster_scope_lock ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 FD: 16 BD: 7 +...: &service->lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->krc.lock FD: 24 BD: 49 +.+.: &pnettable->lock ->&rq->__lock FD: 1 BD: 49 +.+.: smc_ib_devices.mutex FD: 1 BD: 1 +.+.: smc_wr_rx_hash_lock FD: 1 BD: 1 +.+.: v9fs_trans_lock FD: 1 BD: 5 +...: &this->receive_lock FD: 1 BD: 1 +...: lowpan_nhc_lock FD: 238 BD: 7 +.+.: ovs_mutex ->(work_completion)(&data->gc_work) ->nf_ct_proto_mutex ->&obj_hash[i].lock ->pool_lock#2 ->net_rwsem ->ovs_mutex.wait_lock ->&rq->__lock ->quarantine_lock FD: 227 BD: 9 +.+.: defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 227 BD: 9 +.+.: defrag6_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 34 BD: 2 +.+.: drain_vmap_work ->vmap_purge_lock FD: 1 BD: 6 +.+.: ima_keys_lock FD: 52 BD: 88 +.+.: scomp_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock FD: 1 BD: 697 +.+.: &mm->page_table_lock FD: 32 BD: 698 +.+.: ptlock_ptr(ptdesc)#2 ->lock#4 ->key ->&____s->seqcount ->lock#5 ->&folio_wait_table[i] FD: 261 BD: 5 +.+.: k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->local_mutex ->&local->services_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&rx->incoming_lock ->&obj_hash[i].lock ->&rxnet->conn_lock ->&call->waitq ->(rxrpc_call_limiter).lock ->&rx->recvmsg_lock ->&rx->call_lock ->&rxnet->call_lock ->(&call->timer) ->&base->lock ->&list->lock#17 ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock ->&meta->lock ->&n->list_lock ->rcu_node_0 FD: 1 BD: 6 +...: k-slock-AF_RXRPC FD: 248 BD: 6 +.+.: &rxnet->local_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->crngs.lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&____s->seqcount ->&c->lock ->&dir->lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->cpu_hotplug_lock ->&rq->__lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&x->wait#21 ->&cfs_rq->removed.lock ->&n->list_lock ->&____s->seqcount#2 ->stock_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 6 BD: 52 +...: &table->hash[i].lock ->k-clock-AF_INET6 ->&table->hash2[i].lock ->k-clock-AF_INET ->clock-AF_INET ->clock-AF_INET6 FD: 1 BD: 53 +...: &table->hash2[i].lock FD: 229 BD: 2 +.+.: netstamp_work ->cpu_hotplug_lock FD: 25 BD: 7 ....: &x->wait#21 ->&p->pi_lock FD: 1 BD: 6 +.+.: &local->services_lock FD: 1 BD: 8 +.+.: &rxnet->conn_lock FD: 1 BD: 6 ....: &call->waitq FD: 1 BD: 6 +.+.: &rx->call_lock FD: 1 BD: 6 +.+.: &rxnet->call_lock FD: 57 BD: 1 +.+.: init_user_ns.keyring_sem ->key_user_lock ->root_key_user.lock ->fs_reclaim ->pool_lock#2 ->crngs.lock ->key_serial_lock ->key_construction_mutex ->&type->lock_class ->keyring_serialise_link_lock FD: 1 BD: 5 +.+.: root_key_user.lock FD: 1 BD: 8 +.+.: keyring_name_lock FD: 1 BD: 1 +.+.: template_list FD: 1 BD: 1 +.+.: idr_lock FD: 49 BD: 1 +.+.: ima_extend_list_mutex ->fs_reclaim ->pool_lock#2 FD: 30 BD: 5 +.-.: (&rxnet->peer_keepalive_timer) FD: 16 BD: 5 +.+.: (wq_completion)krxrpcd ->(work_completion)(&rxnet->peer_keepalive_work) ->(work_completion)(&rxnet->service_conn_reaper) FD: 13 BD: 6 +.+.: (work_completion)(&rxnet->peer_keepalive_work) ->&rxnet->peer_hash_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 7 +.+.: &rxnet->peer_hash_lock FD: 25 BD: 4 +.+.: deferred_probe_work ->deferred_probe_mutex FD: 24 BD: 45 +.+.: &(&net->nexthop.notifier_chain)->rwsem ->&rq->__lock FD: 35 BD: 46 +.+.: k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->&table->hash[i].lock ->&obj_hash[i].lock ->k-clock-AF_INET ->&rq->__lock FD: 15 BD: 47 +...: k-slock-AF_INET#2 ->pool_lock#2 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 53 ++..: k-clock-AF_INET FD: 1 BD: 1 ....: power_off_handler_list.lock FD: 850 BD: 2 +.+.: reg_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 45 +...: reg_pending_beacons_lock FD: 861 BD: 2 +.+.: (work_completion)(&fw_work->work) ->fs_reclaim ->pool_lock#2 ->&fw_cache.lock ->tk_core.seq.seqcount ->async_lock ->init_task.alloc_lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->(console_sem).lock ->&rq->__lock ->&cfs_rq->removed.lock ->umhelper_sem ->fw_lock ->rtnl_mutex FD: 2 BD: 3 +.+.: &fw_cache.lock ->pool_lock#2 FD: 2 BD: 414 +.+.: &____s->seqcount#6 ->&____s->seqcount#6/1 FD: 1 BD: 1 +.+.: detector_work FD: 1 BD: 1 +.+.: acpi_gpio_deferred_req_irqs_lock FD: 1 BD: 1 ....: enable_lock FD: 1 BD: 1 +.+.: gpd_list_lock FD: 3 BD: 4 +.+.: subsys mutex#73 ->&k->k_lock FD: 2 BD: 18 +.+.: fw_lock ->&x->wait#22 FD: 1 BD: 19 ....: &x->wait#22 FD: 1 BD: 1 +.+.: cdev_lock FD: 255 BD: 3 +.+.: &tty->legacy_mutex ->&tty->read_wait ->&tty->write_wait ->&tty->ldisc_sem ->&tty->files_lock ->&port->lock ->&port->mutex ->&port_lock_key ->&f->f_lock ->&obj_hash[i].lock ->pool_lock#2 ->tasklist_lock FD: 1 BD: 8 ....: &tty->read_wait FD: 25 BD: 219 ....: &tty->write_wait ->&p->pi_lock FD: 243 BD: 4 ++++: &tty->ldisc_sem ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&tty->write_wait ->&tty->read_wait ->&tty->termios_rwsem ->&mm->mmap_lock ->&port_lock_key ->&port->lock ->&tty->flow.lock ->&ldata->atomic_read_lock FD: 158 BD: 7 ++++: &tty->termios_rwsem ->&port->mutex ->&tty->write_wait ->&tty->read_wait ->&ldata->output_lock ->&port_lock_key FD: 1 BD: 6 +.+.: &tty->files_lock FD: 1 BD: 219 ....: &port->lock FD: 25 BD: 412 ....: &wq#2 ->&p->pi_lock FD: 155 BD: 1 +.+.: &type->s_umount_key#24/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->&obj_hash[i].lock ->&____s->seqcount ->&wq->mutex ->&c->lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->wq_pool_mutex ->mmu_notifier_invalidate_range_start ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&xa->xa_lock#9 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->wq_mayday_lock ->&sbi->old_work_lock ->(work_completion)(&(&sbi->old_work)->work) ->&x->wait#23 FD: 1 BD: 11 +.+.: &bdev->bd_holder_lock FD: 25 BD: 105 -.-.: bit_wait_table + i ->&p->pi_lock FD: 140 BD: 79 ++++: &vma->vm_lock->lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->ptlock_ptr(ptdesc)#2 ->mmu_notifier_invalidate_range_start ->mapping.invalidate_lock ->rcu_node_0 ->&rq->__lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&lruvec->lru_lock ->remove_cache_srcu ->&c->lock ->&rcu_state.gp_wq ->&folio_wait_table[i] ->&mm->page_table_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->batched_entropy_u8.lock ->kfence_freelist_lock ->&sem->wait_lock ->&cfs_rq->removed.lock ->stock_lock ->&xa->xa_lock#9 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->tk_core.seq.seqcount ->mount_lock ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->&n->list_lock FD: 1 BD: 2 +.+.: &sbi->old_work_lock FD: 1 BD: 2 +.+.: (work_completion)(&(&sbi->old_work)->work) FD: 1 BD: 9 ....: &x->wait#23 FD: 154 BD: 1 +.+.: &type->s_umount_key#25/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&c->lock ->crypto_alg_sem ->lock#2 ->&x->wait#23 FD: 16 BD: 51 +.+.: (work_completion)(work) ->lock#4 ->lock#5 FD: 334 BD: 1 +.+.: &type->s_umount_key#26/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&xa->xa_lock#9 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->crypto_alg_sem ->pool_lock#2 ->percpu_counters_lock ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_mutex_key#8 ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&journal->j_state_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&journal->j_wait_done_commit ->&p->alloc_lock ->cpu_hotplug_lock ->wq_pool_mutex ->&ei->i_es_lock ->ext4_grpinfo_slab_create_mutex ->&s->s_inode_list_lock ->ext4_li_mtx ->lock ->&root->kernfs_rwsem ->(console_sem).lock ->&dentry->d_lock FD: 17 BD: 93 +.+.: &bgl->locks[i].lock ->&sbi->s_md_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&ei->i_prealloc_lock ->&base->lock FD: 46 BD: 285 +.+.: &sb->s_type->i_lock_key#22 ->&dentry->d_lock ->&lru->node[i].lock ->&xa->xa_lock#9 FD: 182 BD: 5 ++++: &sb->s_type->i_mutex_key#8 ->&ei->i_es_lock ->&ei->i_data_sem ->&ei->xattr_sem ->tk_core.seq.seqcount ->&mm->mmap_lock ->fs_reclaim ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&mapping->i_private_lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&c->lock ->rcu_node_0 ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->mapping.invalidate_lock ->&meta->lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->swap_cgroup_mutex ->&fq->mq_flush_lock ->&x->wait#26 ->&base->lock ->(&timer.timer) ->&((cluster_info + ci)->lock)/1 ->swapon_mutex ->proc_poll_wait.lock ->&dentry->d_lock ->smack_known_lock ->&n->list_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&rcu_state.expedited_wq ->&sem->wait_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&p->pi_lock FD: 19 BD: 88 ++++: &ei->i_es_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&sbi->s_es_lock ->&obj_hash[i].lock ->key#2 ->key#7 ->key#8 ->key#9 ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 97 BD: 87 ++++: &ei->i_data_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&ei->i_es_lock ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->&ei->i_prealloc_lock ->&sb->s_type->i_lock_key#22 ->&(ei->i_block_reservation_lock) ->&lg->lg_mutex ->&mapping->i_private_lock ->&ei->i_raw_lock ->&ret->b_state_lock ->&rq->__lock ->&wb->list_lock ->lock#4 ->&journal->j_revoke_lock ->key#14 ->&sbi->s_md_lock ->key#3 ->&n->list_lock ->&____s->seqcount#2 ->rcu_node_0 ->&cfs_rq->removed.lock ->remove_cache_srcu ->&bgl->locks[i].lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->quarantine_lock ->&wb->work_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock ->bit_wait_table + i ->&pa->pa_lock#2 ->&xa->xa_lock#9 ->&sem->wait_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&p->pi_lock FD: 1 BD: 89 +.+.: &sbi->s_es_lock FD: 62 BD: 92 ++++: &journal->j_state_lock ->&journal->j_wait_done_commit ->&journal->j_wait_commit ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->&journal->j_wait_updates ->&journal->j_wait_transaction_locked ->&journal->j_list_lock ->&journal->j_wait_reserved FD: 25 BD: 93 ....: &journal->j_wait_done_commit ->&p->pi_lock FD: 25 BD: 93 ....: &journal->j_wait_commit ->&p->pi_lock FD: 102 BD: 2 +.+.: ext4_grpinfo_slab_create_mutex ->slab_mutex FD: 53 BD: 3 +.+.: ext4_li_mtx ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->batched_entropy_u16.lock ->&eli->li_list_mtx ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: &rs->lock FD: 165 BD: 6 ++++: &type->i_mutex_dir_key#3 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&obj_hash[i].lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->&ei->xattr_sem ->namespace_sem ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->key ->pcpu_lock ->percpu_counters_lock ->tomoyo_ss ->&s->s_inode_list_lock ->jbd2_handle ->&mm->mmap_lock ->&n->list_lock ->remove_cache_srcu ->stock_lock ->rcu_node_0 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&meta->lock ->&base->lock ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->&dentry->d_lock/1 ->&sem->wait_lock FD: 72 BD: 87 ++++: &ei->xattr_sem ->mmu_notifier_invalidate_range_start ->lock#4 ->&mapping->i_private_lock ->pool_lock#2 ->&ret->b_state_lock ->&journal->j_revoke_lock ->tk_core.seq.seqcount ->&ei->i_raw_lock ->&____s->seqcount ->&xa->xa_lock#9 ->&c->lock ->&rq->__lock ->stock_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->bit_wait_table + i FD: 39 BD: 47 +.+.: rcu_state.barrier_mutex ->rcu_state.barrier_lock ->&x->wait#24 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.barrier_mutex.wait_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->pool_lock FD: 25 BD: 48 ..-.: &x->wait#24 ->&p->pi_lock FD: 45 BD: 696 ++++: &anon_vma->rwsem ->&mm->page_table_lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&rq->__lock ->mmu_notifier_invalidate_range_start ->ptlock_ptr(ptdesc)#2 ->&sem->wait_lock ->quarantine_lock ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->&base->lock ->&rcu_state.expedited_wq ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 190 BD: 1 +.+.: &sig->cred_guard_mutex ->fs_reclaim ->pool_lock#2 ->init_fs.lock ->&p->pi_lock ->mapping.invalidate_lock ->&folio_wait_table[i] ->&rq->__lock ->tomoyo_ss ->binfmt_lock ->init_binfmt_misc.entries_lock ->&dentry->d_lock ->&type->i_mutex_dir_key#3 ->&sb->s_type->i_lock_key#22 ->&obj_hash[i].lock ->&ei->xattr_sem ->&tsk->futex_exit_mutex ->&sig->exec_update_lock ->&fs->lock ->lock#4 ->&sb->s_type->i_mutex_key#8 ->&p->alloc_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&mm->mmap_lock ->&stopper->lock ->&stop_pi_lock ->&x->wait#8 ->&n->list_lock ->quarantine_lock ->remove_cache_srcu ->key#5 ->rcu_node_0 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq FD: 134 BD: 81 ++++: mapping.invalidate_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&ei->i_es_lock ->&ei->i_data_sem ->pool_lock#2 ->tk_core.seq.seqcount ->&c->lock ->&folio_wait_table[i] ->&rq->__lock ->&obj_hash[i].lock ->&mapping->i_mmap_rwsem ->&journal->j_state_lock ->jbd2_handle FD: 1 BD: 4 ++++: init_binfmt_misc.entries_lock FD: 173 BD: 2 +.+.: &sig->exec_update_lock ->&p->alloc_lock ->&sighand->siglock ->&newf->file_lock ->batched_entropy_u64.lock ->&mm->mmap_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&rq->__lock ->pool_lock ->&cfs_rq->removed.lock ->stock_lock ->quarantine_lock FD: 3 BD: 12 ..-.: batched_entropy_u16.lock ->crngs.lock FD: 1 BD: 699 +.+.: ptlock_ptr(ptdesc)#2/1 FD: 1 BD: 89 ....: key#2 FD: 1 BD: 17 ..-.: task_group_lock FD: 95 BD: 1 +.+.: &type->s_umount_key#27/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#9 ->&dentry->d_lock FD: 37 BD: 117 +.+.: &sb->s_type->i_lock_key#23 ->&dentry->d_lock ->&lru->node[i].lock FD: 83 BD: 3 ++++: &sb->s_type->i_mutex_key#9 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->rename_lock.seqcount ->proc_subdir_lock ->&c->lock ->&p->alloc_lock ->&pid->lock ->sysctl_lock ->namespace_sem ->tomoyo_ss ->&n->list_lock ->&rq->__lock ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 ->remove_cache_srcu ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 896 BD: 2 +.+.: &p->lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->file_systems_lock ->namespace_sem ->&c->lock ->&____s->seqcount ->&of->mutex ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&rq->__lock ->module_mutex ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 95 BD: 1 +.+.: &type->s_umount_key#28/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#24 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock FD: 37 BD: 286 +.+.: &sb->s_type->i_lock_key#24 ->&dentry->d_lock ->&lru->node[i].lock ->&p->pi_lock FD: 172 BD: 3 ++++: &type->i_mutex_dir_key#4 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&root->kernfs_rwsem ->mmu_notifier_invalidate_range_start ->iattr_mutex ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#24 ->namespace_sem ->&c->lock ->&____s->seqcount ->&mm->mmap_lock ->vmap_area_lock ->tk_core.seq.seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->remove_cache_srcu ->&rq->__lock ->rcu_node_0 ->rename_lock ->&sem->wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->&rcu_state.gp_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock ->&xa->xa_lock#5 FD: 49 BD: 225 +.+.: iattr_mutex ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->tk_core.seq.seqcount ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 61 BD: 1 +.+.: &type->s_umount_key#29 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 25 BD: 129 ....: &x->wait#25 ->&p->pi_lock FD: 38 BD: 8 +.+.: &net->unx.table.locks[i] ->&net->unx.table.locks[i]/1 FD: 944 BD: 1 +.+.: &sb->s_type->i_mutex_key#10 ->&net->unx.table.locks[i] ->&u->lock ->&u->peer_wait ->rlock-AF_UNIX ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->sk_lock-AF_INET ->slock-AF_INET#2 ->clock-AF_INET ->rcu_node_0 ->&rq->__lock ->nl_table_lock ->nl_table_wait.lock ->clock-AF_NETLINK ->&nlk->wait ->(netlink_chain).rwsem ->sk_lock-AF_INET6 ->slock-AF_INET6 ->clock-AF_INET6 ->&table->hash[i].lock ->&net->packet.sklist_lock ->&po->bind_lock ->sk_lock-AF_PACKET ->slock-AF_PACKET ->fanout_mutex ->&x->wait#2 ->clock-AF_PACKET ->pcpu_lock ->quarantine_lock ->cb_lock ->genl_sk_destructing_waitq.lock ->sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->hci_dev_list_lock ->stock_lock ->sk_lock-AF_VSOCK ->slock-AF_VSOCK ->sk_lock-AF_KCM ->slock-AF_KCM ->&mux->lock ->(work_completion)(&kcm->tx_work) ->&mux->rx_lock ->&knet->mutex ->sk_lock-AF_PHONET ->slock-AF_PHONET ->clock-AF_PHONET ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&cfs_rq->removed.lock ->&rnp->exp_lock ->&rnp->exp_wq[2] ->rcu_state.exp_mutex ->&rnp->exp_wq[3] ->&base->lock ->&rcu_state.expedited_wq ->&meta->lock ->kfence_freelist_lock ->&net->ipv4.ra_mutex ->&hashinfo->lock ->clock-AF_ROSE ->sk_lock-AF_ROSE ->slock-AF_ROSE ->&hashinfo->lock#2 ->clock-AF_NETROM ->sk_lock-AF_NETROM ->slock-AF_NETROM ->key ->percpu_counters_lock ->&____s->seqcount ->sk_lock-AF_INET/1 ->&net->sctp.addr_wq_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock FD: 46 BD: 3 +.+.: &u->lock ->clock-AF_UNIX ->&u->lock/1 ->rlock-AF_UNIX ->&u->peer_wait ->&sk->sk_peer_lock ->&ei->socket.wq.wait FD: 1 BD: 4 +...: clock-AF_UNIX FD: 25 BD: 4 +.+.: &u->peer_wait ->&p->pi_lock FD: 1 BD: 4 +.+.: rlock-AF_UNIX FD: 30 BD: 1 ..-.: &(&ipvs->defense_work)->timer FD: 30 BD: 1 ..-.: &(&gc_work->dwork)->timer FD: 28 BD: 6 +.+.: (work_completion)(&(&ipvs->defense_work)->work) ->&s->s_inode_list_lock ->&ipvs->dropentry_lock ->&ipvs->droppacket_lock ->&ipvs->securetcp_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 37 BD: 2 +.+.: (work_completion)(&(&gc_work->dwork)->work) ->rcu_node_0 ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 1 BD: 218 ...-: &____s->seqcount#7 FD: 1 BD: 7 +...: &ipvs->dropentry_lock FD: 1 BD: 7 +...: &ipvs->droppacket_lock FD: 1 BD: 7 +...: &ipvs->securetcp_lock FD: 183 BD: 1 .+.+: sb_writers#3 ->mount_lock ->tk_core.seq.seqcount ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&wb->work_lock ->&type->i_mutex_dir_key#3 ->&type->i_mutex_dir_key#3/1 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&zone->lock ->&xa->xa_lock#9 ->lock#4 ->&mapping->i_private_lock ->&base->lock ->bit_wait_table + i ->&rq->__lock ->remove_cache_srcu ->&sb->s_type->i_mutex_key#8 ->tomoyo_ss ->&n->list_lock ->&s->s_inode_list_lock ->sb_internal ->inode_hash_lock ->&fsnotify_mark_srcu ->&mm->mmap_lock ->fs_reclaim ->&____s->seqcount#2 ->quarantine_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&sem->wait_lock ->&p->pi_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&p->alloc_lock ->&f->f_lock FD: 164 BD: 1 .+.+: sb_writers#4 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#23 ->&wb->list_lock ->&sb->s_type->i_mutex_key#9 ->sysctl_lock ->&dentry->d_lock ->tomoyo_ss ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&obj_hash[i].lock ->oom_adj_mutex ->&p->pi_lock ->&c->lock ->&____s->seqcount#10 ->&(&net->ipv4.ping_group_range.lock)->lock ->remove_cache_srcu ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->oom_adj_mutex.wait_lock ->&rcu_state.expedited_wq FD: 1 BD: 4 +.+.: &pid->lock FD: 167 BD: 1 ++++: &type->s_umount_key#30 ->&lru->node[i].lock ->&dentry->d_lock ->&sb->s_type->i_lock_key#22 ->&sbi->s_writepages_rwsem ->&sem->waiters ->&rsp->gp_wait ->&journal->j_state_lock ->&p->alloc_lock ->(work_completion)(&sbi->s_sb_upd_work) ->key#3 ->key#4 ->&sbi->s_error_lock ->mmu_notifier_invalidate_range_start ->batched_entropy_u8.lock ->kfence_freelist_lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->bit_wait_table + i ->&rq->__lock ->ext4_li_mtx ->(console_sem).lock ->mount_lock ->&eli->li_list_mtx ->&wb->list_lock FD: 130 BD: 4 ++++: &sbi->s_writepages_rwsem ->&rsp->gp_wait ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->lock#4 ->lock#5 ->&journal->j_state_lock ->jbd2_handle ->tk_core.seq.seqcount ->&xa->xa_lock#9 ->&base->lock ->rcu_node_0 ->&rq_wait->wait ->&__ctx->lock ->&n->list_lock ->&____s->seqcount#2 ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock FD: 1 BD: 2 ....: &sem->waiters FD: 128 BD: 3 +.+.: (work_completion)(&sbi->s_sb_upd_work) ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock FD: 1 BD: 92 ....: key#3 FD: 1 BD: 91 ....: key#4 FD: 1 BD: 91 +.+.: &sbi->s_error_lock FD: 4 BD: 4 +.+.: &eli->li_list_mtx ->&obj_hash[i].lock ->pool_lock#2 FD: 127 BD: 90 ++++: jbd2_handle ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&ret->b_state_lock ->&journal->j_revoke_lock ->&ei->i_raw_lock ->&journal->j_wait_updates ->&mapping->i_private_lock ->&meta_group_info[i]->alloc_sem ->tk_core.seq.seqcount ->inode_hash_lock ->batched_entropy_u32.lock ->&ei->xattr_sem ->&obj_hash[i].lock ->&ei->i_es_lock ->&dentry->d_lock ->smack_known_lock ->&sb->s_type->i_lock_key#22 ->&rq->__lock ->&journal->j_state_lock ->bit_wait_table + i ->lock#4 ->lock#5 ->&ei->i_data_sem ->&xa->xa_lock#9 ->&sbi->s_orphan_lock ->&journal->j_list_lock ->key#3 ->key#4 ->&sbi->s_error_lock ->&n->list_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->stock_lock ->&____s->seqcount#2 ->&ei->i_prealloc_lock ->&bgl->locks[i].lock ->remove_cache_srcu ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock ->smack_known_lock.wait_lock ->&p->pi_lock ->quarantine_lock ->&lock->wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&sem->wait_lock ->&rq_wait->wait ->&__ctx->lock FD: 57 BD: 95 +.+.: &ret->b_state_lock ->&journal->j_list_lock ->&obj_hash[i].lock FD: 56 BD: 98 +.+.: &journal->j_list_lock ->&sb->s_type->i_lock_key#3 ->&wb->list_lock ->&wb->work_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 94 +.+.: &journal->j_revoke_lock FD: 1 BD: 94 +.+.: &ei->i_raw_lock FD: 25 BD: 93 ....: &journal->j_wait_updates ->&p->pi_lock FD: 12 BD: 306 -...: &wb->work_lock ->&obj_hash[i].lock ->&base->lock FD: 52 BD: 91 ++++: &meta_group_info[i]->alloc_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&rq->__lock ->&base->lock ->&c->lock ->&____s->seqcount ->&x->wait#26 ->&__ctx->lock ->rcu_node_0 ->(&timer.timer) ->&fq->mq_flush_lock ->&bgl->locks[i].lock ->&cfs_rq->removed.lock FD: 132 BD: 3 .+.+: sb_internal ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&base->lock ->rcu_node_0 FD: 2 BD: 95 ++++: &ei->i_prealloc_lock ->&pa->pa_lock#2 FD: 26 BD: 1 .+.+: file_rwsem ->&ctx->flc_lock ->&rq->__lock FD: 2 BD: 2 +.+.: &ctx->flc_lock ->&fll->lock FD: 1 BD: 3 +.+.: &fll->lock FD: 159 BD: 2 +.+.: &type->i_mutex_dir_key#3/1 ->rename_lock.seqcount ->&dentry->d_lock ->fs_reclaim ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->pool_lock#2 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&obj_hash[i].lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->&ei->xattr_sem ->tomoyo_ss ->&s->s_inode_list_lock ->jbd2_handle ->&c->lock ->&sb->s_type->i_mutex_key#8 ->rcu_node_0 ->&n->list_lock ->&xa->xa_lock#5 ->stock_lock ->&____s->seqcount#2 ->&fsnotify_mark_srcu ->&type->i_mutex_dir_key#3 ->&wb->list_lock ->sb_internal ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->quarantine_lock ->remove_cache_srcu ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock ->key ->pcpu_lock ->percpu_counters_lock ->&sem->wait_lock ->&p->pi_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#31/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#11 ->&dentry->d_lock FD: 37 BD: 3 +.+.: &sb->s_type->i_lock_key#25 ->&dentry->d_lock FD: 62 BD: 2 +.+.: &sb->s_type->i_mutex_key#11 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount FD: 62 BD: 1 +.+.: &type->s_umount_key#32 ->sb_lock ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&lru->node[i].lock ->&obj_hash[i].lock FD: 39 BD: 1 +.+.: &type->s_umount_key#33 ->sb_lock ->&dentry->d_lock FD: 39 BD: 1 +.+.: &type->s_umount_key#34 ->sb_lock ->&dentry->d_lock FD: 75 BD: 1 +.+.: &type->s_umount_key#35/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&dentry->d_lock FD: 37 BD: 4 +.+.: &sb->s_type->i_lock_key#26 ->&dentry->d_lock FD: 39 BD: 1 +.+.: &type->s_umount_key#36 ->sb_lock ->&dentry->d_lock FD: 1 BD: 1 +.+.: redirect_lock FD: 240 BD: 1 +.+.: &tty->atomic_write_lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&tty->termios_rwsem ->&tty->files_lock FD: 36 BD: 8 +.+.: &ldata->output_lock ->&port_lock_key ->&rq->__lock FD: 76 BD: 1 +.+.: &type->s_umount_key#37/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->&c->lock ->&____s->seqcount ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#27 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->fuse_mutex ->&dentry->d_lock FD: 37 BD: 112 +.+.: &sb->s_type->i_lock_key#27 ->&dentry->d_lock FD: 1 BD: 2 +.+.: fuse_mutex FD: 81 BD: 1 +.+.: &type->s_umount_key#38/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#28 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->pstore_sb_lock ->&sb->s_type->i_mutex_key#12 ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#28 ->&dentry->d_lock FD: 57 BD: 4 +.+.: &sb->s_type->i_mutex_key#12 ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&psinfo->read_mutex ->&obj_hash[i].lock FD: 56 BD: 5 +.+.: &psinfo->read_mutex ->(efivars_lock).lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->(efi_runtime_lock).lock ->&obj_hash[i].lock ->&x->wait#12 ->&rq->__lock FD: 79 BD: 1 +.+.: &type->s_umount_key#39/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#29 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->bpf_preload_lock ->&dentry->d_lock FD: 37 BD: 2 +.+.: &sb->s_type->i_lock_key#29 ->&dentry->d_lock FD: 56 BD: 2 +.+.: bpf_preload_lock ->(kmod_concurrent_max).lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->key ->pcpu_lock ->percpu_counters_lock ->running_helpers_waitq.lock FD: 25 BD: 1 ++++: uts_sem ->hostname_poll.wait.lock ->&rq->__lock FD: 156 BD: 3 ++++: &type->i_mutex_dir_key#5 ->fs_reclaim ->&dentry->d_lock ->rename_lock.seqcount ->&c->lock ->&____s->seqcount ->tomoyo_ss ->&sbinfo->stat_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&xattrs->lock ->&obj_hash[i].lock ->&simple_offset_xa_lock ->smack_known_lock ->remove_cache_srcu ->&n->list_lock ->&rq->__lock ->&sem->wait_lock ->rcu_node_0 ->&p->pi_lock ->smack_known_lock.wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock ->&mm->mmap_lock ->vmap_area_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock ->&meta->lock FD: 169 BD: 2 .+.+: sb_writers#5 ->mount_lock ->&type->i_mutex_dir_key#5 ->&type->i_mutex_dir_key#5/1 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key ->&wb->list_lock ->&sb->s_type->i_mutex_key#13 ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->&s->s_inode_list_lock ->&info->lock ->&obj_hash[i].lock ->pool_lock#2 ->&sbinfo->stat_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->tomoyo_ss ->&xattrs->lock ->fs_reclaim ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&dentry->d_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&rcu_state.expedited_wq FD: 94 BD: 3 +.+.: &type->i_mutex_dir_key#5/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->tomoyo_ss ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&xattrs->lock ->&obj_hash[i].lock ->&simple_offset_xa_lock ->smack_known_lock ->&u->bindlock ->pool_lock#2 ->&sb->s_type->i_mutex_key#13/4 ->&sem->wait_lock ->&rq->__lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->&sb->s_type->i_mutex_key#13 ->&fsnotify_mark_srcu ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&info->lock ->&xa->xa_lock#9 ->smack_known_lock.wait_lock ->&p->pi_lock ->remove_cache_srcu ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->quarantine_lock ->key#6 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 26 BD: 1 +.-.: (&cb->timer) ->&obj_hash[i].lock ->&base->lock ->tk_core.seq.seqcount ->&rq_wait->wait FD: 2 BD: 7 +.+.: &f->f_lock ->fasync_lock FD: 1 BD: 2 ....: hostname_poll.wait.lock FD: 907 BD: 1 +.+.: &f->f_pos_lock ->&type->i_mutex_dir_key#3 ->&mm->mmap_lock ->&type->i_mutex_dir_key#4 ->sb_writers#5 ->&type->i_mutex_dir_key#5 ->&p->lock ->sysctl_lock ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->sb_writers#7 ->&sb->s_type->i_mutex_key#17 ->rcu_node_0 FD: 144 BD: 78 +.+.: &mm->mmap_lock/1 ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&vma->vm_lock->lock ->&mapping->i_mmap_rwsem ->&anon_vma->rwsem ->mmu_notifier_invalidate_range_start ->&mm->page_table_lock ->ptlock_ptr(ptdesc)#2 ->remove_cache_srcu ->&n->list_lock ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->stock_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock ->&____s->seqcount#2 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->key#21 ->key ->pcpu_lock ->percpu_counters_lock FD: 158 BD: 4 +.+.: &sb->s_type->i_mutex_key#13 ->&xattrs->lock ->tk_core.seq.seqcount ->&mm->mmap_lock ->fs_reclaim ->&____s->seqcount ->&xa->xa_lock#9 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->&wb->list_lock ->key#6 ->&rq->__lock ->&sb->s_type->i_mutex_key#13/4 ->rcu_node_0 ->&simple_offset_xa_lock ->&dentry->d_lock ->&mapping->i_mmap_rwsem ->lock#5 ->&lruvec->lru_lock ->&obj_hash[i].lock ->tomoyo_ss ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&cfs_rq->removed.lock ->pool_lock#2 ->&rcu_state.expedited_wq FD: 1 BD: 2 ....: key#5 FD: 64 BD: 6 +.+.: &u->bindlock ->&net->unx.table.locks[i] ->&bsd_socket_locks[i] ->fs_reclaim ->pool_lock#2 ->batched_entropy_u32.lock ->&net->unx.table.locks[i]/1 FD: 37 BD: 9 +.+.: &net->unx.table.locks[i]/1 ->&dentry->d_lock FD: 1 BD: 7 +.+.: &bsd_socket_locks[i] FD: 173 BD: 1 +.+.: &u->iolock ->rlock-AF_UNIX ->&mm->mmap_lock ->&obj_hash[i].lock ->pool_lock#2 ->&u->peer_wait ->&rq->__lock ->quarantine_lock ->&u->lock ->&meta->lock ->kfence_freelist_lock ->&dir->lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&base->lock ->&rcu_state.expedited_wq FD: 27 BD: 372 ..-.: &ei->socket.wq.wait ->&p->pi_lock ->&ep->lock FD: 39 BD: 4 +.+.: &u->lock/1 ->&sk->sk_peer_lock ->&dentry->d_lock ->&sk->sk_peer_lock/1 FD: 166 BD: 1 +.+.: &pipe->mutex/1 ->&pipe->rd_wait ->&pipe->wr_wait ->fs_reclaim ->&____s->seqcount ->&mm->mmap_lock ->&rq->__lock ->&lock->wait_lock ->&obj_hash[i].lock ->rcu_node_0 ->stock_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock#2 FD: 27 BD: 4 ....: &pipe->rd_wait ->&p->pi_lock ->&ep->lock FD: 1 BD: 5 ....: key#6 FD: 27 BD: 4 ....: &pipe->wr_wait ->&p->pi_lock ->&ep->lock FD: 41 BD: 1 .+.+: sb_writers#6 ->tk_core.seq.seqcount ->mount_lock ->&rq->__lock FD: 170 BD: 1 +.+.: sk_lock-AF_NETLINK ->slock-AF_NETLINK ->&mm->mmap_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&c->lock ->batched_entropy_u32.lock ->vmap_purge_lock ->&fp->aux->used_maps_mutex ->&rq->__lock FD: 1 BD: 2 +...: slock-AF_NETLINK FD: 2 BD: 6 +.+.: &sk->sk_peer_lock ->&sk->sk_peer_lock/1 FD: 1 BD: 1 ....: &rs->lock#2 FD: 1 BD: 89 ....: key#7 FD: 1 BD: 89 ....: key#8 FD: 1 BD: 89 ....: key#9 FD: 49 BD: 2 +.+.: oom_adj_mutex ->&p->alloc_lock ->&rq->__lock ->rcu_node_0 ->oom_adj_mutex.wait_lock FD: 76 BD: 1 +.+.: &group->mark_mutex ->&fsnotify_mark_srcu ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->lock ->ucounts_lock ->&mark->lock ->&conn->lock ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->&rq->__lock ->remove_cache_srcu ->&____s->seqcount#2 ->&lock->wait_lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 10 BD: 217 +.+.: &group->inotify_data.idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 3 BD: 2 +.+.: &mark->lock ->&fsnotify_mark_srcu ->&conn->lock FD: 1 BD: 7 +.+.: &conn->lock FD: 181 BD: 2 +.+.: &ep->mtx ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&f->f_lock ->&ei->socket.wq.wait ->&ep->lock ->&group->notification_waitq ->&group->notification_lock ->&sighand->signalfd_wqh ->&sighand->siglock ->&mm->mmap_lock ->&rq->__lock ->&pipe->rd_wait ->key#10 ->&obj_hash[i].lock ->&lock->wait_lock ->sysctl_lock ->&pipe->wr_wait ->remove_cache_srcu ->rcu_node_0 ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock ->&____s->seqcount#2 ->stock_lock ->&rcu_state.expedited_wq FD: 182 BD: 1 +.+.: epnested_mutex ->&ep->mtx FD: 26 BD: 386 ...-: &ep->lock ->&ep->wq FD: 27 BD: 7 ....: &group->notification_waitq ->&p->pi_lock ->&ep->lock FD: 1 BD: 7 +.+.: &group->notification_lock FD: 27 BD: 98 ....: &sighand->signalfd_wqh ->&ep->lock ->&p->pi_lock FD: 12 BD: 5 +.-.: (&net->can.stattimer) ->&obj_hash[i].lock ->&base->lock FD: 887 BD: 2 .+.+: sb_writers#7 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->&wb->list_lock ->&type->i_mutex_dir_key#4 ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&root->kernfs_iattr_rwsem ->&dentry->d_lock ->tomoyo_ss ->iattr_mutex ->&sb->s_type->i_mutex_key#14 ->&xattrs->lock ->&____s->seqcount#2 ->&____s->seqcount ->quarantine_lock FD: 30 BD: 1 ..-.: &(&krcp->monitor_work)->timer FD: 30 BD: 1 ..-.: &(&tbl->managed_work)->timer FD: 32 BD: 2 +.+.: (work_completion)(&(&krcp->monitor_work)->work) ->krc.lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 3 BD: 8 +.+.: swap_lock ->&p->lock#2 FD: 66 BD: 1 .+.+: kn->active ->fs_reclaim ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&k->list_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->quarantine_lock FD: 49 BD: 66 +.+.: &kernfs_locks->open_file_mutex[count] ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->remove_cache_srcu ->&rq->__lock ->&____s->seqcount#2 ->rcu_node_0 ->&lock->wait_lock ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock FD: 890 BD: 6 +.+.: &of->mutex ->pool_lock#2 ->&obj_hash[i].lock ->&rq->__lock ->cgroup_mutex ->&cfs_rq->removed.lock ->&root->deactivate_waitq FD: 1 BD: 367 ..-.: rlock-AF_NETLINK FD: 25 BD: 387 ..-.: &ep->wq ->&p->pi_lock FD: 1 BD: 6 ....: &nlk->wait FD: 65 BD: 1 .+.+: kn->active#2 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->remove_cache_srcu ->quarantine_lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 31 BD: 2 +.+.: (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->krc.lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 18 BD: 1 +.-.: (&vblank->disable_timer) ->&dev->vbl_lock FD: 65 BD: 1 .+.+: kn->active#3 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->&____s->seqcount ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&n->list_lock ->quarantine_lock ->&rq->__lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 30 BD: 1 +.-.: (&q->timeout) FD: 34 BD: 1 +.+.: (wq_completion)kblockd ->(work_completion)(&q->timeout_work) ->(work_completion)(&(&hctx->run_work)->work) ->(work_completion)(&(&q->requeue_work)->work) FD: 2 BD: 2 +.+.: (work_completion)(&q->timeout_work) ->&tags->lock FD: 124 BD: 9 ++++: kn->active#4 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->remove_cache_srcu ->&n->list_lock ->&device->physical_node_lock ->udc_lock ->fw_lock ->quarantine_lock ->&rfkill->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rq->__lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&root->deactivate_waitq ->&meta->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->udc_lock.wait_lock ->&rcu_state.expedited_wq ->&lock->wait_lock ->&base->lock FD: 1 BD: 7 +.+.: &sk->sk_peer_lock/1 FD: 1 BD: 718 ....: &sem->wait_lock FD: 25 BD: 93 -...: &x->wait#26 ->&p->pi_lock FD: 1 BD: 99 +.+.: &__ctx->lock FD: 34 BD: 92 -.-.: &fq->mq_flush_lock ->tk_core.seq.seqcount ->&q->requeue_lock ->&obj_hash[i].lock ->&x->wait#26 ->bit_wait_table + i ->quarantine_lock FD: 1 BD: 96 -.-.: &q->requeue_lock FD: 30 BD: 1 ..-.: &(&ovs_net->masks_rebalance)->timer FD: 239 BD: 6 +.+.: (work_completion)(&(&ovs_net->masks_rebalance)->work) ->ovs_mutex ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->ovs_mutex.wait_lock ->&p->pi_lock FD: 55 BD: 1 .+.+: kn->active#5 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->param_lock ->pool_lock#2 ->&on->poll ->&rq->__lock FD: 51 BD: 3 +.+.: &sb->s_type->i_mutex_key#14 ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 47 +.+.: disk_events_mutex FD: 63 BD: 5 +.+.: &sb->s_type->i_mutex_key#13/4 ->&dentry->d_lock ->&simple_offset_xa_lock ->fs_reclaim ->tk_core.seq.seqcount ->rename_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 29 BD: 412 +.+.: &dentry->d_lock/2 ->&dentry->d_lock/3 FD: 28 BD: 413 +.+.: &dentry->d_lock/3 ->&____s->seqcount#6 ->&wq FD: 1 BD: 415 +.+.: &____s->seqcount#6/1 FD: 51 BD: 1 .+.+: kn->active#6 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 51 BD: 1 .+.+: kn->active#7 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#8 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#9 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#10 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#11 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->&rq->__lock ->remove_cache_srcu FD: 53 BD: 1 .+.+: kn->active#12 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 50 BD: 1 .+.+: kn->active#13 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 ->&rq->__lock ->&n->list_lock ->remove_cache_srcu FD: 53 BD: 1 .+.+: kn->active#14 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 47 BD: 1 .+.+: kn->active#15 ->fs_reclaim ->remove_cache_srcu ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&____s->seqcount FD: 51 BD: 1 .+.+: kn->active#16 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 51 BD: 1 .+.+: kn->active#17 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount FD: 51 BD: 1 .+.+: kn->active#18 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 51 BD: 1 .+.+: kn->active#19 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 48 BD: 1 .+.+: kn->active#20 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->remove_cache_srcu FD: 53 BD: 1 .+.+: kn->active#21 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock FD: 1 BD: 1 +.+.: &sb->s_type->i_mutex_key#15 FD: 46 BD: 1 .+.+: mapping.invalidate_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->pool_lock#2 ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&rq->__lock FD: 59 BD: 1 .+.+: kn->active#22 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&dev->power.lock ->pci_lock FD: 51 BD: 1 .+.+: kn->active#23 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 51 BD: 1 .+.+: kn->active#24 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#25 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 70 BD: 49 +.-.: slock-AF_INET/1 ->tk_core.seq.seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&____s->seqcount#2 ->&zone->lock ->&sctp_ep_hashtable[i].lock ->clock-AF_INET FD: 103 BD: 45 +.+.: devnet_rename_sem ->(console_sem).lock ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->&root->kernfs_rwsem ->&c->lock ->kernfs_rename_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount ->&____s->seqcount#2 ->&sem->wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->remove_cache_srcu FD: 51 BD: 1 .+.+: kn->active#26 ->&rq->__lock ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 224 ....: kernfs_rename_lock FD: 49 BD: 1 .+.+: kn->active#27 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount FD: 51 BD: 1 .+.+: kn->active#28 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 51 BD: 1 .+.+: kn->active#29 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&rq->__lock FD: 54 BD: 3 ++++: kn->active#30 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->remove_cache_srcu ->&n->list_lock ->&____s->seqcount ->&lock->wait_lock ->&rq->__lock ->&p->pi_lock ->&root->deactivate_waitq FD: 51 BD: 1 .+.+: kn->active#31 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->&n->list_lock FD: 49 BD: 1 .+.+: kn->active#32 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount FD: 51 BD: 1 .+.+: kn->active#33 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 1 +.+.: &evdev->client_lock FD: 27 BD: 1 +.+.: &evdev->mutex ->&dev->mutex#2 FD: 51 BD: 1 .+.+: kn->active#34 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 49 +.+.: &nft_net->commit_mutex FD: 1 BD: 49 ....: target_list_lock FD: 209 BD: 2 +.+.: sk_lock-AF_INET ->slock-AF_INET#2 ->&table->hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&icsk->icsk_accept_queue.rskq_lock ->clock-AF_INET ->&obj_hash[i].lock ->&base->lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&mm->mmap_lock ->tk_core.seq.seqcount ->&sd->defer_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->quarantine_lock ->once_mutex ->batched_entropy_u32.lock ->batched_entropy_u16.lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->&meta->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 73 BD: 7 +.-.: slock-AF_INET#2 ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&hashinfo->ehash_locks[i] ->tk_core.seq.seqcount ->(&req->rsk_timer) ->&base->lock ->&icsk->icsk_accept_queue.rskq_lock ->pool_lock#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->&sk->sk_lock.wq FD: 1 BD: 68 ++..: clock-AF_INET FD: 51 BD: 1 .+.+: kn->active#35 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 96 +.+.: smack_known_lock.wait_lock FD: 51 BD: 1 .+.+: kn->active#36 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#37 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 3 ....: key#10 FD: 53 BD: 2 +.+.: &dev_instance->mutex ->fs_reclaim ->pool_lock#2 ->vicodec_core:1851:(hdl)->_lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->remove_cache_srcu ->quarantine_lock FD: 24 BD: 3 +.+.: vicodec_core:1851:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 1 BD: 5 ....: &vdev->fh_lock FD: 58 BD: 1 +.+.: &mdev->req_queue_mutex ->&dev_instance->mutex ->&vdev->fh_lock ->&mdev->graph_mutex ->vicodec_core:1851:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 ->vim2m:1183:(hdl)->_lock ->&dev->dev_mutex ->&dev->mutex#3 ->quarantine_lock ->&rq->__lock FD: 1 BD: 4 ....: &m2m_dev->job_spinlock FD: 1 BD: 6 ....: &q->done_wq FD: 1 BD: 6 +.+.: &q->mmap_lock FD: 54 BD: 2 +.+.: &dev->dev_mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->&____s->seqcount ->vim2m:1183:(hdl)->_lock ->&obj_hash[i].lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock ->&n->list_lock FD: 4 BD: 3 +.+.: vim2m:1183:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 27 BD: 2 +.+.: &dev->mutex#3 ->&vdev->fh_lock ->&q->done_wq ->&q->mmap_lock ->&rq->__lock FD: 51 BD: 1 .+.+: kn->active#38 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount FD: 1 BD: 1 +.+.: fh->state->lock FD: 3 BD: 1 +.+.: &vcapture->lock ->&q->done_wq ->&q->mmap_lock FD: 24 BD: 11 +.+.: &lo->lo_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 44 BD: 11 +.+.: &nbd->config_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&bdev->bd_size_lock ->&q->queue_lock ->&ACCESS_PRIVATE(sdp, lock) ->set->srcu ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#3 FD: 30 BD: 5 ....: &ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&obj_hash[i].lock ->&base->lock FD: 2 BD: 10 +.+.: &new->lock ->&mtdblk->cache_mutex FD: 1 BD: 11 +.+.: &mtdblk->cache_mutex FD: 51 BD: 1 .+.+: kn->active#39 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 162 BD: 1 +.+.: &mtd->master.chrdev_lock ->&mm->mmap_lock FD: 51 BD: 1 .+.+: kn->active#40 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 4 +.+.: destroy_lock FD: 30 BD: 1 ..-.: fs/notify/mark.c:89 FD: 61 BD: 2 +.+.: connector_reaper_work ->&rq->__lock ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&x->wait#3 ->pool_lock#2 ->rcu_node_0 ->&cfs_rq->removed.lock ->&base->lock ->quarantine_lock ->pool_lock FD: 61 BD: 2 +.+.: (reaper_work).work ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&x->wait#3 ->&rq->__lock ->pool_lock#2 ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&base->lock ->pool_lock ->&meta->lock ->kfence_freelist_lock FD: 30 BD: 1 ..-.: &(&wb->dwork)->timer FD: 134 BD: 1 +.+.: (wq_completion)writeback ->(work_completion)(&(&wb->dwork)->work) ->(work_completion)(&(&wb->bw_dwork)->work) FD: 132 BD: 2 +.+.: (work_completion)(&(&wb->dwork)->work) ->&wb->work_lock ->&wb->list_lock ->&p->sequence ->key#11 ->&pl->lock ->&rq->__lock FD: 2 BD: 4 +.-.: &p->sequence ->key#13 FD: 1 BD: 308 -...: key#11 FD: 1 BD: 1 ....: (&journal->j_commit_timer) FD: 72 BD: 1 +.+.: &journal->j_checkpoint_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&journal->j_state_lock FD: 25 BD: 93 ....: &journal->j_wait_transaction_locked ->&p->pi_lock FD: 1 BD: 100 -...: &memcg->move_lock FD: 1 BD: 90 +.+.: &sbi->s_md_lock FD: 1 BD: 1 ....: &journal->j_fc_wait FD: 1 BD: 1 +.+.: &journal->j_history_lock FD: 10 BD: 46 +...: fib_info_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 74 BD: 46 +...: &net->sctp.local_addr_lock ->&net->sctp.addr_wq_lock FD: 73 BD: 48 +.-.: &net->sctp.addr_wq_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->k-slock-AF_INET6/1 ->slock-AF_INET/1 FD: 52 BD: 48 +.+.: (work_completion)(&ht->run_work) ->&ht->mutex FD: 51 BD: 49 +.+.: &ht->mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->batched_entropy_u32.lock ->rhashtable_bucket ->&ht->lock ->remove_cache_srcu ->&rq->__lock ->&____s->seqcount ->&____s->seqcount#2 ->&n->list_lock ->&obj_hash[i].lock ->&meta->lock ->kfence_freelist_lock ->batched_entropy_u8.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->quarantine_lock FD: 1 BD: 99 ....: rhashtable_bucket/1 FD: 4 BD: 50 +.+.: &ht->lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +...: clock-AF_NETLINK FD: 1 BD: 45 +...: _xmit_LOOPBACK FD: 24 BD: 51 .+.+: netpoll_srcu ->&rq->__lock FD: 13 BD: 56 +.-.: &in_dev->mc_tomb_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->quarantine_lock FD: 16 BD: 52 +.-.: &im->lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 FD: 1 BD: 51 +.+.: cbs_list_lock FD: 13 BD: 49 +...: &net->ipv6.addrconf_hash_lock ->&obj_hash[i].lock FD: 31 BD: 80 +...: &ifa->lock ->batched_entropy_u32.lock ->crngs.lock ->&obj_hash[i].lock ->&base->lock FD: 41 BD: 81 +...: &tb->tb6_lock ->&net->ipv6.fib6_walker_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->rlock-AF_NETLINK ->rt6_exception_lock ->&data->fib_event_queue_lock ->quarantine_lock ->&n->list_lock ->&____s->seqcount#2 ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 1 BD: 82 ++..: &net->ipv6.fib6_walker_lock FD: 206 BD: 2 +.+.: sk_lock-AF_INET6 ->slock-AF_INET6 ->&table->hash[i].lock ->batched_entropy_u32.lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->fs_reclaim ->&mm->mmap_lock ->once_lock ->rcu_node_0 ->&rq->__lock ->clock-AF_INET6 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 20 BD: 3 +...: slock-AF_INET6 ->&obj_hash[i].lock ->&tcp_hashinfo.bhash[i].lock FD: 1 BD: 68 ++..: clock-AF_INET6 FD: 1 BD: 1 +.+.: userns_state_mutex FD: 2 BD: 1 +.+.: sk_lock-AF_UNIX ->slock-AF_UNIX FD: 1 BD: 2 +...: slock-AF_UNIX FD: 51 BD: 1 .+.+: kn->active#41 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 163 BD: 45 ++++: dev_addr_sem ->net_rwsem ->&tn->lock ->&sdata->sec_mtx ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->&tbl->lock ->&pn->hash_lock ->&obj_hash[i].lock ->input_pool.lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&br->lock ->&n->list_lock ->&cfs_rq->removed.lock ->_xmit_ETHER ->&hard_iface->bat_iv.ogm_buff_mutex ->rcu_node_0 ->remove_cache_srcu ->&____s->seqcount#2 ->quarantine_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->team->team_lock_key#34 ->team->team_lock_key#33 ->team->team_lock_key#35 ->&base->lock ->team->team_lock_key#36 ->team->team_lock_key#38 FD: 860 BD: 3 +.+.: nlk_cb_mutex-GENERIC ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->rtnl_mutex ->&rdev->wiphy.mtx ->rlock-AF_NETLINK ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->&devlink->lock_key#33 ->&devlink->lock_key#34 ->&devlink->lock_key#35 ->&devlink->lock_key#36 ->&devlink->lock_key#38 FD: 18 BD: 49 +...: &rdev->bss_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->quarantine_lock ->&base->lock FD: 74 BD: 1 +.-.: (&net->sctp.addr_wq_timer) ->&net->sctp.addr_wq_lock FD: 1 BD: 46 +.+.: napi_hash_lock FD: 12 BD: 45 ++..: lapb_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock FD: 37 BD: 2 +.+.: (work_completion)(&aux->work) ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->vmap_purge_lock ->pool_lock#2 ->pcpu_lock ->stock_lock ->quarantine_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&base->lock FD: 4 BD: 45 ++.-: x25_neigh_list_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 45 +...: _xmit_SLIP FD: 13 BD: 1 +.-.: (&eql->timer) ->&eql->queue.lock ->&obj_hash[i].lock ->&base->lock FD: 4 BD: 48 +.-.: &eql->queue.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 51 BD: 87 +.+.: xps_map_mutex ->fs_reclaim ->pool_lock#2 ->jump_label_mutex ->&rq->__lock FD: 1 BD: 49 +.+.: &data->mutex FD: 17 BD: 66 +...: &local->filter_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 30 BD: 67 ..-.: &rdev->wiphy_work_lock FD: 400 BD: 7 +.+.: (work_completion)(&rdev->wiphy_work) ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 45 +...: _xmit_VOID FD: 1 BD: 211 +.-.: &ul->lock FD: 1 BD: 45 +...: _xmit_X25 FD: 4 BD: 46 +...: &lapbeth->up_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 60 BD: 46 +.-.: &lapb->lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&list->lock#20 FD: 2 BD: 88 +.+.: &(ei->i_block_reservation_lock) ->key#14 FD: 31 BD: 2 +.+.: (work_completion)(&work->work) ->devices_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 850 BD: 2 +.+.: (work_completion)(&(&ifa->dad_work)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 8 +.+.: fasync_lock FD: 1 BD: 99 ....: &tty->ctrl.lock FD: 1 BD: 1 +.+.: &buf->lock FD: 1 BD: 8 ....: &tty->flow.lock FD: 1 BD: 82 +.-.: rt6_exception_lock FD: 73 BD: 45 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->_xmit_ETHER#2 ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&sch->q.lock ->&r->producer_lock#3 ->_xmit_SLIP#2 ->_xmit_NETROM ->quarantine_lock FD: 57 BD: 71 +.-.: _xmit_ETHER#2 ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock FD: 160 BD: 5 +.+.: &ldata->atomic_read_lock ->&tty->termios_rwsem ->(work_completion)(&buf->work) ->&rq->__lock FD: 30 BD: 1 +.-.: &tx->clean_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 30 BD: 1 ..-.: &(&idev->mc_dad_work)->timer FD: 135 BD: 1 +.+.: (wq_completion)mld ->(work_completion)(&(&idev->mc_dad_work)->work) ->(work_completion)(&(&idev->mc_ifc_work)->work) FD: 133 BD: 2 +.+.: (work_completion)(&(&idev->mc_dad_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 1 BD: 50 ....: class FD: 1 BD: 50 ....: (&tbl->proxy_timer) FD: 30 BD: 1 ..-.: &(&idev->mc_ifc_work)->timer FD: 133 BD: 2 +.+.: (work_completion)(&(&idev->mc_ifc_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 15 BD: 54 +.-.: &ul->lock#2 ->pool_lock#2 ->&dir->lock#2 ->&c->lock ->&n->list_lock ->&____s->seqcount FD: 14 BD: 204 ++--: &n->lock ->&____s->seqcount#8 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->pool_lock#2 ->&(&n->ha_lock)->lock FD: 1 BD: 206 +.--: &____s->seqcount#8 FD: 30 BD: 1 ..-.: net/core/link_watch.c:31 FD: 80 BD: 46 +.-.: &dev->tx_global_lock ->_xmit_ETHER#2 ->&obj_hash[i].lock ->&base->lock ->_xmit_NETROM ->_xmit_NONE#2 ->_xmit_TUNNEL6#2 ->_xmit_SIT#2 ->_xmit_TUNNEL#2 ->_xmit_IPGRE#2 ->&qdisc_xmit_lock_key ->&qdisc_xmit_lock_key#2 ->&vlan_netdev_xmit_lock_key ->&batadv_netdev_xmit_lock_key ->&qdisc_xmit_lock_key#3 ->&qdisc_xmit_lock_key#4 ->_xmit_LOOPBACK#2 FD: 1 BD: 47 +...: &sch->q.lock FD: 1 BD: 203 +.-.: lock#8 FD: 1 BD: 203 ..-.: id_table_lock FD: 1 BD: 216 ...-: &____s->seqcount#9 FD: 31 BD: 2 +.+.: (work_completion)(&w->work)#2 ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->nf_conntrack_mutex FD: 1 BD: 2 +.+.: &net->packet.sklist_lock FD: 181 BD: 2 +.+.: sk_lock-AF_PACKET ->slock-AF_PACKET ->&po->bind_lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&mm->mmap_lock ->fs_reclaim ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->vmap_purge_lock ->&fp->aux->used_maps_mutex ->&c->lock ->&n->list_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->&cfs_rq->removed.lock FD: 1 BD: 3 +...: slock-AF_PACKET FD: 13 BD: 3 +.+.: &po->bind_lock ->ptype_lock ->pool_lock#2 ->&dir->lock#2 FD: 1 BD: 2 +.+.: fanout_mutex FD: 1 BD: 1 +.-.: rlock-AF_PACKET FD: 1 BD: 1 +...: wlock-AF_PACKET FD: 30 BD: 1 ..-.: &(&ifa->dad_work)->timer FD: 1 BD: 2 +...: clock-AF_PACKET FD: 30 BD: 1 ..-.: drivers/base/dd.c:321 FD: 37 BD: 2 +.+.: (deferred_probe_timeout_work).work ->device_links_lock ->deferred_probe_mutex ->deferred_probe_work ->&x->wait#10 ->&rq->__lock ->&obj_hash[i].lock FD: 66 BD: 1 .+.+: &type->s_umount_key#40 ->&sb->s_type->i_lock_key#3 ->mmu_notifier_invalidate_range_start ->batched_entropy_u8.lock ->kfence_freelist_lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->lock#4 ->lock#5 ->&wb->list_lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&____s->seqcount#2 ->&n->list_lock FD: 1 BD: 308 -...: &s->s_inode_wblist_lock FD: 1 BD: 309 -...: key#12 FD: 79 BD: 88 +.+.: &lg->lg_mutex ->&ei->i_prealloc_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->lock#4 ->&mapping->i_private_lock ->&ret->b_state_lock ->&journal->j_revoke_lock ->&pa->pa_lock ->&lg->lg_prealloc_lock ->bit_wait_table + i ->&__ctx->lock ->&obj_hash[i].lock ->&rq->__lock FD: 1 BD: 89 +.+.: &pa->pa_lock FD: 1 BD: 89 +.+.: &lg->lg_prealloc_lock FD: 30 BD: 3 -.-.: &ei->i_completed_io_lock FD: 133 BD: 1 +.+.: (wq_completion)ext4-rsv-conversion ->(work_completion)(&ei->i_rsv_conversion_work) FD: 132 BD: 2 +.+.: (work_completion)(&ei->i_rsv_conversion_work) ->&ei->i_completed_io_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&ext4__ioend_wq[i] ->&ret->b_uptodate_lock ->&folio_wait_table[i] ->&rq->__lock ->rcu_node_0 ->quarantine_lock ->&lruvec->lru_lock ->&rcu_state.expedited_wq FD: 1 BD: 93 ....: &journal->j_wait_reserved FD: 1 BD: 3 ....: &ext4__ioend_wq[i] FD: 30 BD: 1 ..-.: &(&wb->bw_dwork)->timer FD: 55 BD: 2 +.+.: (work_completion)(&(&wb->bw_dwork)->work) ->&wb->list_lock FD: 61 BD: 1 +.-.: (&lapb->t1timer) ->&lapb->lock FD: 81 BD: 1 +.-.: (&dev->watchdog_timer) ->&dev->tx_global_lock FD: 14 BD: 1 +.-.: (&dom->period_timer) ->key#13 ->&p->sequence ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 310 -.-.: key#13 FD: 1 BD: 56 +.-.: &ct->lock FD: 5 BD: 201 +.-.: &nf_conntrack_locks[i] ->&nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 4 BD: 202 +.-.: &nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 1 BD: 66 +.-.: &hashinfo->ehash_locks[i] FD: 2 BD: 205 +.-.: &(&n->ha_lock)->lock ->&____s->seqcount#8 FD: 1 BD: 8 ..-.: (&req->rsk_timer) FD: 1 BD: 8 +.-.: &icsk->icsk_accept_queue.rskq_lock FD: 1 BD: 3 +.-.: &sd->defer_lock FD: 74 BD: 1 +.-.: (&icsk->icsk_delack_timer) ->slock-AF_INET#2 FD: 74 BD: 1 +.-.: (&icsk->icsk_retransmit_timer) ->slock-AF_INET#2 FD: 2 BD: 50 +.-.: tcp_metrics_lock ->pool_lock#2 FD: 1 BD: 89 ....: key#14 FD: 2 BD: 308 ....: &pl->lock ->key#12 FD: 69 BD: 91 +.+.: &sbi->s_orphan_lock ->mmu_notifier_invalidate_range_start ->&ei->i_raw_lock ->&rq->__lock ->&mapping->i_private_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&lock->wait_lock ->&rcu_state.expedited_wq ->&ret->b_state_lock FD: 51 BD: 1 .+.+: kn->active#42 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 1 +.+.: &futex_queues[i].lock FD: 1 BD: 4 ....: &on->poll FD: 1 BD: 3 +.+.: module_mutex FD: 24 BD: 3 +.+.: once_mutex ->crngs.lock ->&rq->__lock FD: 95 BD: 1 +.+.: &type->s_umount_key#41/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#30 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock FD: 37 BD: 265 +.+.: &sb->s_type->i_lock_key#30 ->&dentry->d_lock FD: 882 BD: 1 .+.+: sb_writers#8 ->mount_lock ->&type->i_mutex_dir_key#6 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#6/1 ->&c->lock ->&n->list_lock FD: 66 BD: 2 ++++: &type->i_mutex_dir_key#6 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#30 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->pool_lock#2 ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock FD: 51 BD: 1 ++++: kn->active#43 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 FD: 95 BD: 1 +.+.: &type->s_umount_key#42/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock ->&c->lock ->&n->list_lock FD: 37 BD: 265 +.+.: &sb->s_type->i_lock_key#31 ->&dentry->d_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#43 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->percpu_ref_switch_lock ->&root->kernfs_supers_rwsem ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#31 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->inode_hash_lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 1 BD: 1 ..-.: percpu_ref_switch_waitq.lock FD: 861 BD: 2 +.+.: (work_completion)(&cgrp->bpf.release_work) ->cgroup_mutex ->cgroup_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 16 +.+.: cgroup_mutex.wait_lock FD: 866 BD: 1 +.+.: (wq_completion)cgroup_destroy ->(work_completion)(&css->destroy_work) ->(work_completion)(&(&css->destroy_rwork)->work) FD: 859 BD: 2 +.+.: (work_completion)(&css->destroy_work) ->cgroup_mutex ->&obj_hash[i].lock ->pool_lock#2 ->cgroup_mutex.wait_lock ->&p->pi_lock FD: 864 BD: 2 +.+.: (work_completion)(&(&css->destroy_rwork)->work) ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 ->&cgrp->pidlist_mutex ->(wq_completion)cgroup_pidlist_destroy ->&wq->mutex ->(work_completion)(&cgrp->release_agent_work) ->cgroup_mutex ->cgroup_rstat_lock ->pcpu_lock ->&root->kernfs_rwsem ->kernfs_idr_lock ->krc.lock ->cgroup_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 3 +.+.: &cgrp->pidlist_mutex FD: 1 BD: 3 +.+.: (wq_completion)cgroup_pidlist_destroy FD: 1 BD: 3 +.+.: (work_completion)(&cgrp->release_agent_work) FD: 882 BD: 1 .+.+: sb_writers#9 ->mount_lock ->&type->i_mutex_dir_key#7 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#7/1 ->&c->lock ->remove_cache_srcu ->&n->list_lock FD: 65 BD: 2 ++++: &type->i_mutex_dir_key#7 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->pool_lock#2 ->&c->lock ->&rq->__lock ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 16 +.+.: &dom->lock FD: 51 BD: 1 .+.+: kn->active#44 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 229 BD: 1 .+.+: kn->active#45 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->cpu_hotplug_lock FD: 176 BD: 1 .+.+: sb_writers#10 ->&mm->mmap_lock ->&attr->mutex ->mount_lock ->&sb->s_type->i_mutex_key#3 FD: 162 BD: 2 +.+.: &attr->mutex ->&mm->mmap_lock FD: 39 BD: 1 +.+.: &type->s_umount_key#44 ->sb_lock ->&dentry->d_lock FD: 79 BD: 2 +.+.: &sb->s_type->i_mutex_key#16 ->namespace_sem ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&c->lock ->&____s->seqcount ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->init_binfmt_misc.entries_lock FD: 170 BD: 1 .+.+: sb_writers#11 ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&sb->s_type->i_mutex_key#16 FD: 1 BD: 45 +.+.: &wpan_dev->association_lock FD: 1 BD: 46 +...: &pn->hash_lock FD: 42 BD: 1 +...: &net->ipv6.fib6_gc_lock ->&obj_hash[i].lock FD: 1 BD: 45 +...: _xmit_IEEE802154 FD: 1 BD: 3 +.+.: &xa->xa_lock#16 FD: 1 BD: 6 ....: genl_sk_destructing_waitq.lock FD: 1 BD: 3 +...: &rdev->beacon_registrations_lock FD: 1 BD: 50 +...: &rdev->mgmt_registrations_lock FD: 1 BD: 50 +...: &wdev->pmsr_lock FD: 1 BD: 46 +.+.: reg_indoor_lock FD: 146 BD: 2 +.+.: (work_completion)(&w->w) ->nfc_devlist_mutex ->&obj_hash[i].lock ->pool_lock#2 ->nfc_devlist_mutex.wait_lock ->&p->pi_lock ->&meta->lock ->kfence_freelist_lock ->&base->lock ->&rq->__lock FD: 1 BD: 5 +.+.: &genl_data->genl_data_mutex FD: 49 BD: 6 +.+.: swap_cgroup_mutex ->fs_reclaim ->&____s->seqcount FD: 1 BD: 6 +.+.: &((cluster_info + ci)->lock)/1 FD: 67 BD: 6 +.+.: swapon_mutex ->fs_reclaim ->pool_lock#2 ->swap_lock ->percpu_ref_switch_lock ->(console_sem).lock FD: 2 BD: 9 +.+.: &p->lock#2 ->swap_avail_lock FD: 1 BD: 10 +.+.: swap_avail_lock FD: 1 BD: 6 ....: proc_poll_wait.lock FD: 232 BD: 1 +.+.: swap_slots_cache_enable_mutex ->cpu_hotplug_lock ->swap_lock FD: 1 BD: 89 +.+.: swap_slots_cache_mutex FD: 1 BD: 130 ....: &newf->resize_wait FD: 13 BD: 134 ..-.: &kcov->lock ->kcov_remote_lock FD: 102 BD: 1 +.+.: pid_caches_mutex ->slab_mutex FD: 39 BD: 1 +.+.: &type->s_umount_key#45 ->&rq->__lock ->sb_lock ->&dentry->d_lock FD: 167 BD: 2 ++++: &sb->s_type->i_mutex_key#17 ->&rq->__lock ->namespace_sem ->&mm->mmap_lock ->vmap_area_lock ->&dentry->d_lock ->tk_core.seq.seqcount FD: 1 BD: 54 ++++: hci_sk_list.lock FD: 1 BD: 1 +.+.: (work_completion)(&(&data->open_timeout)->work) FD: 255 BD: 1 +.+.: &data->open_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&obj_hash[i].lock ->&x->wait#9 ->hci_index_ida.xa_lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->wq_pool_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&n->list_lock ->&rq->__lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#74 ->&dev->devres_lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rfkill->lock ->hci_dev_list_lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->(pm_chain_head).rwsem ->&list->lock#7 ->&data->read_wait ->&sem->wait_lock ->uevent_sock_mutex.wait_lock ->&____s->seqcount#2 FD: 1 BD: 2 ....: hci_index_ida.xa_lock FD: 26 BD: 51 +.+.: subsys mutex#74 ->&rq->__lock ->&k->k_lock FD: 1 BD: 28 ++++: hci_dev_list_lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#46 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 182 BD: 23 +.+.: (work_completion)(&hdev->power_on) ->&hdev->req_lock ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&rq->__lock FD: 181 BD: 25 +.+.: &hdev->req_lock ->&obj_hash[i].lock ->&c->lock ->pool_lock#2 ->&list->lock#5 ->&list->lock#6 ->&hdev->req_wait_q ->&base->lock ->&rq->__lock ->(&timer.timer) ->&____s->seqcount ->tk_core.seq.seqcount ->hci_sk_list.lock ->pool_lock ->(work_completion)(&(&hdev->interleave_scan)->work) ->hci_dev_list_lock ->(work_completion)(&hdev->tx_work) ->(work_completion)(&hdev->rx_work) ->&wq->mutex ->&hdev->lock ->&list->lock#7 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->&n->list_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->(wq_completion)hci1#8 ->(wq_completion)hci2#8 ->(wq_completion)hci4#6 ->(wq_completion)hci1#10 ->(wq_completion)hci2#10 ->(wq_completion)hci0#8 ->(wq_completion)hci2#12 ->(wq_completion)hci4#8 ->(wq_completion)hci5#4 ->(wq_completion)hci1#12 ->(wq_completion)hci4#10 ->(wq_completion)hci1#14 ->(wq_completion)hci3#8 ->(wq_completion)hci2#14 ->(wq_completion)hci0#10 ->(wq_completion)hci4#12 ->(wq_completion)hci0#12 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 26 ....: &list->lock#5 FD: 1 BD: 52 ....: &list->lock#6 FD: 25 BD: 49 ....: &hdev->req_wait_q ->&p->pi_lock FD: 1 BD: 51 ....: &list->lock#7 FD: 25 BD: 51 ....: &data->read_wait ->&p->pi_lock FD: 50 BD: 2 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->sock_cookie_ida.xa_lock ->&p->alloc_lock ->&c->lock ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->clock-AF_BLUETOOTH ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&rq->__lock FD: 1 BD: 3 +...: slock-AF_BLUETOOTH-BTPROTO_HCI FD: 1 BD: 3 ....: sock_cookie_ida.xa_lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#23 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 50 BD: 48 +.+.: (work_completion)(&hdev->cmd_work) ->&list->lock#6 ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->&list->lock#7 ->&data->read_wait ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->remove_cache_srcu ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 151 BD: 48 +.+.: (work_completion)(&hdev->rx_work) ->&list->lock#6 ->lock#6 ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&hdev->lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&obj_hash[i].lock ->&hdev->req_wait_q ->&rq->__lock ->&base->lock ->&c->lock ->&n->list_lock ->&meta->lock ->kfence_freelist_lock ->chan_list_lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->rcu_node_0 ->quarantine_lock ->remove_cache_srcu ->&rcu_state.expedited_wq FD: 138 BD: 49 +.+.: &hdev->lock ->&xa->xa_lock#17 ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#9 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&k->k_lock ->subsys mutex#74 ->&list->lock#6 ->&hdev->unregister_lock ->hci_cb_list_lock ->&base->lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->pool_lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->uevent_sock_mutex.wait_lock ->&n->list_lock ->(work_completion)(&(&conn->disc_work)->work) ->(work_completion)(&(&conn->auto_accept_work)->work) ->(work_completion)(&(&conn->idle_work)->work) ->&x->wait#2 ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount#2 ->hci_cb_list_lock.wait_lock ->remove_cache_srcu ->quarantine_lock FD: 12 BD: 50 ....: &xa->xa_lock#17 ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 54 BD: 51 +.+.: &hdev->unregister_lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->&hdev->cmd_sync_work_lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock FD: 1 BD: 52 +.+.: &hdev->cmd_sync_work_lock FD: 182 BD: 23 +.+.: (work_completion)(&hdev->cmd_sync_work) ->&hdev->cmd_sync_work_lock ->&hdev->req_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 51 +.+.: &conn->ident_lock FD: 1 BD: 52 ....: &list->lock#8 FD: 1 BD: 53 +.+.: &conn->chan_lock FD: 29 BD: 48 +.+.: (work_completion)(&hdev->tx_work) ->&list->lock#8 ->tk_core.seq.seqcount ->&list->lock#7 ->&data->read_wait ->&list->lock#6 FD: 2 BD: 48 +.+.: (work_completion)(&conn->pending_rx_work) ->&list->lock#9 FD: 1 BD: 49 ....: &list->lock#9 FD: 1 BD: 3 +...: clock-AF_BLUETOOTH FD: 1 BD: 1 +.+.: &undo_list->lock FD: 1 BD: 45 +...: &nr_netdev_addr_lock_key FD: 1 BD: 45 +...: listen_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#358 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1990 FD: 2 BD: 6 +.+.: rdma_nets.xa_lock ->pool_lock#2 FD: 1 BD: 45 +...: &bat_priv->forw_bcast_list_lock FD: 1 BD: 3 +.+.: &____s->seqcount#10 FD: 2 BD: 2 +.+.: &(&net->ipv4.ping_group_range.lock)->lock ->&____s->seqcount#10 FD: 2 BD: 45 +.+.: &r->consumer_lock ->&r->producer_lock FD: 1 BD: 55 +...: &r->producer_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#291 ->&rq->__lock FD: 10 BD: 53 +...: &bridge_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 35 BD: 48 +.-.: &br->hash_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 1 BD: 45 +.+.: j1939_netdev_lock FD: 1 BD: 189 +.+.: uevent_sock_mutex.wait_lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#64 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#16 ->(work_completion)(&(&slave->notify_work)->work) FD: 8 BD: 53 +...: &dev_addr_list_lock_key#2 ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1788 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1788 FD: 9 BD: 45 +...: &bat_priv->tvlv.handler_list_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 14 BD: 52 +...: &bat_priv->tvlv.container_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->quarantine_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 8 BD: 53 +...: &batadv_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 10 BD: 54 +...: &bat_priv->softif_vlan_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 17 BD: 53 +...: key#15 ->&bat_priv->softif_vlan_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 4 BD: 52 +...: &bat_priv->tt.changes_list_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 30 BD: 1 ..-.: &(&bat_priv->nc.work)->timer FD: 60 BD: 1 +.+.: (wq_completion)bat_events ->(work_completion)(&(&bat_priv->nc.work)->work) ->(work_completion)(&(&bat_priv->mcast.work)->work) ->(work_completion)(&(&bat_priv->orig_work)->work) ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->(work_completion)(&(&bat_priv->tt.work)->work) ->(work_completion)(&(&bat_priv->dat.work)->work) ->(work_completion)(&(&bat_priv->bla.work)->work) ->&rq->__lock ->(work_completion)(&barr->work) FD: 30 BD: 6 +.+.: (work_completion)(&(&bat_priv->nc.work)->work) ->key#16 ->key#17 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->pool_lock#2 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 7 +...: key#16 FD: 1 BD: 7 +...: key#17 FD: 102 BD: 46 +.+.: init_lock ->slab_mutex ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&base->lock ->crngs.lock FD: 1 BD: 64 +.-.: deferred_lock FD: 850 BD: 2 +.+.: deferred_process_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 50 BD: 47 +.-.: &br->lock ->&br->hash_lock ->lweventlist_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&dir->lock#2 ->deferred_lock ->(console_sem).lock ->&c->lock ->&____s->seqcount ->nl_table_lock ->nl_table_wait.lock ->&br->multicast_lock ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1595 ->&rq->__lock FD: 40 BD: 68 +.+.: (work_completion)(&(&slave->notify_work)->work) ->&obj_hash[i].lock ->&base->lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 46 +.+.: &bond->stats_lock/1 FD: 30 BD: 1 ..-.: &(&slave->notify_work)->timer FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1586 FD: 39 BD: 48 +.+.: &hard_iface->bat_iv.ogm_buff_mutex ->crngs.lock ->pool_lock#2 ->batched_entropy_u8.lock ->&bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&c->lock ->kfence_freelist_lock ->&____s->seqcount ->&bat_priv->tt.commit_lock ->&bat_priv->tvlv.container_list_lock ->&____s->seqcount#2 ->&n->list_lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 14 BD: 49 +...: &bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 45 +...: _xmit_NONE FD: 1 BD: 45 +...: lock#9 FD: 1 BD: 46 ...-: &____s->seqcount#11 FD: 1 BD: 57 +.-.: &hsr->list_lock FD: 30 BD: 1 ..-.: &(&bat_priv->mcast.work)->timer FD: 36 BD: 6 +.+.: (work_completion)(&(&bat_priv->mcast.work)->work) ->pool_lock#2 ->&bat_priv->mcast.mla_lock ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rq->__lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->quarantine_lock ->&rcu_state.expedited_wq FD: 32 BD: 7 +.+.: &bat_priv->mcast.mla_lock ->pool_lock#2 ->key#15 ->&bat_priv->tt.changes_list_lock ->&____s->seqcount ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 FD: 8 BD: 53 +...: &vlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 15 BD: 45 +.-.: (&app->join_timer) ->&app->lock ->&list->lock#10 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 46 +.-.: &app->lock FD: 1 BD: 46 ..-.: &list->lock#10 FD: 9 BD: 53 +...: &macvlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#505 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1706 FD: 8 BD: 53 +...: &dev_addr_list_lock_key#3 ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 1 BD: 45 ....: &xa->xa_lock#18 FD: 16 BD: 53 +...: &dev_addr_list_lock_key#3/1 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&n->list_lock FD: 15 BD: 45 +.-.: (&app->join_timer)#2 ->&app->lock#2 ->&list->lock#11 FD: 13 BD: 47 +.-.: &app->lock#2 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 46 ..-.: &list->lock#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1932 FD: 2 BD: 45 +.+.: &tap_major->minor_lock ->pool_lock#2 FD: 3 BD: 45 +.+.: subsys mutex#75 ->&k->k_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1633 FD: 869 BD: 1 .+.+: kn->active#46 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->&____s->seqcount ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock ->&n->list_lock FD: 869 BD: 9 +.+.: nsim_bus_dev_list_lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->nsim_bus_dev_ids.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->device_links_lock ->&n->list_lock ->deferred_probe_mutex ->&rq->__lock ->subsys mutex#76 ->nsim_bus_dev_list_lock.wait_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount#2 ->&lock->wait_lock ->&p->pi_lock ->&sem->wait_lock ->uevent_sock_mutex.wait_lock ->remove_cache_srcu ->rcu_node_0 FD: 869 BD: 1 .+.+: kn->active#47 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock ->&rq->__lock ->&c->lock ->&n->list_lock FD: 1 BD: 10 ....: nsim_bus_dev_ids.xa_lock FD: 2 BD: 18 +.+.: devlinks.xa_lock ->pool_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1413 FD: 10 BD: 18 +.+.: &xa->xa_lock#19 ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 83 +...: &data->fib_event_queue_lock FD: 1 BD: 18 ....: &(&fn_net->fib_chain)->lock FD: 58 BD: 2 +.+.: (work_completion)(&data->fib_event_work) ->&data->fib_event_queue_lock ->&data->fib_lock ->&rq->__lock FD: 56 BD: 3 +.+.: &data->fib_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->pool_lock ->remove_cache_srcu ->&n->list_lock ->&____s->seqcount ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->quarantine_lock ->&meta->lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1843 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1585 FD: 30 BD: 1 ..-.: &(&nsim_dev->trap_data->trap_report_dw)->timer FD: 32 BD: 2 +.+.: (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 54 BD: 49 +.+.: bpf_devs_lock ->fs_reclaim ->pool_lock#2 ->&rq->__lock ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&____s->seqcount ->remove_cache_srcu FD: 24 BD: 45 +.+.: (work_completion)(&(&devlink_port->type_warn_dw)->work) ->&rq->__lock FD: 1 BD: 45 +...: &devlink_port->type_lock FD: 1 BD: 45 +.+.: &vn->sock_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1671 ->&rq->__lock FD: 1 BD: 10 +.+.: subsys mutex#76 FD: 30 BD: 1 ..-.: &(&hwstats->traffic_dw)->timer FD: 25 BD: 2 +.+.: (work_completion)(&(&hwstats->traffic_dw)->work) ->&hwstats->hwsdev_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 24 BD: 46 +.+.: &hwstats->hwsdev_list_lock ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1835 ->&rq->__lock ->&cfs_rq->removed.lock FD: 30 BD: 1 ..-.: &(&bat_priv->orig_work)->timer FD: 30 BD: 1 ..-.: drivers/net/wireguard/ratelimiter.c:20 FD: 28 BD: 6 +.+.: (work_completion)(&(&bat_priv->orig_work)->work) ->key#18 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->pool_lock#2 ->&rcu_state.expedited_wq FD: 28 BD: 2 +.+.: (gc_work).work ->tk_core.seq.seqcount ->"ratelimiter_table_lock" ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 47 +...: key#18 FD: 1 BD: 3 +.+.: "ratelimiter_table_lock" FD: 17 BD: 20 +.+.: &nsim_trap_data->trap_lock ->pool_lock#2 ->&c->lock ->crngs.lock ->&nsim_dev->fa_cookie_lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->&____s->seqcount ->quarantine_lock ->&____s->seqcount#2 ->&base->lock FD: 1 BD: 21 +...: &nsim_dev->fa_cookie_lock FD: 30 BD: 1 ..-.: &(&dm_bufio_cleanup_old_work)->timer FD: 14 BD: 1 +.+.: (wq_completion)dm_bufio_cache ->(work_completion)(&(&dm_bufio_cleanup_old_work)->work) FD: 13 BD: 2 +.+.: (work_completion)(&(&dm_bufio_cleanup_old_work)->work) ->dm_bufio_clients_lock ->&obj_hash[i].lock ->&base->lock FD: 30 BD: 1 ..-.: &(&forw_packet_aggr->delayed_work)->timer FD: 41 BD: 46 +.+.: (work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->&hard_iface->bat_iv.ogm_buff_mutex ->&bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock ->&n->list_lock ->batched_entropy_u8.lock ->quarantine_lock ->&base->lock ->rcu_node_0 FD: 452 BD: 45 +.+.: &wg->device_update_lock ->&wg->static_identity.lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->pcpu_alloc_mutex ->&handshake->lock ->&obj_hash[i].lock ->pool_lock ->tk_core.seq.seqcount ->&table->lock ->&peer->endpoint_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-slock-AF_INET/1 ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->cpu_hotplug_lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&wg->socket_update_lock ->&rq->__lock ->&list->lock#14 ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&zone->lock ->&rnp->exp_wq[2] ->&____s->seqcount#2 ->&n->list_lock ->&x->wait#2 ->&table->hash[i].lock ->k-clock-AF_INET ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->k-clock-AF_INET6 ->(&peer->timer_retransmit_handshake) ->&base->lock ->(&peer->timer_send_keepalive) ->(&peer->timer_new_handshake) ->(&peer->timer_zero_key_material) ->(&peer->timer_persistent_keepalive) ->(work_completion)(&peer->clear_peer_work) ->&peer->keypairs.keypair_update_lock ->&cfs_rq->removed.lock ->&wq->mutex ->napi_hash_lock ->&table->lock#2 ->wq_pool_mutex ->wq_mayday_lock ->&p->pi_lock ->&x->wait ->pcpu_lock ->&r->consumer_lock#2 ->rcu_state.barrier_mutex ->init_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->quarantine_lock ->remove_cache_srcu ->stock_lock ->key ->percpu_counters_lock ->(wq_completion)wg-crypt-wg2#19 ->(wq_completion)wg-kex-wg2#37 ->(wq_completion)wg-kex-wg2#38 ->(wq_completion)wg-crypt-wg1#19 ->(wq_completion)wg-kex-wg1#37 ->(wq_completion)wg-kex-wg1#38 ->(wq_completion)wg-crypt-wg0#19 ->(wq_completion)wg-kex-wg0#37 ->(wq_completion)wg-kex-wg0#38 ->(wq_completion)wg-crypt-wg2#15 ->(wq_completion)wg-kex-wg2#30 ->(wq_completion)wg-kex-wg2#29 ->(wq_completion)wg-crypt-wg1#15 ->(wq_completion)wg-kex-wg1#29 ->(wq_completion)wg-kex-wg1#30 ->(wq_completion)wg-crypt-wg0#15 ->(wq_completion)wg-kex-wg0#29 ->(wq_completion)wg-kex-wg0#30 ->(wq_completion)wg-crypt-wg2#21 ->(wq_completion)wg-kex-wg2#42 ->(wq_completion)wg-kex-wg2#41 ->(wq_completion)wg-crypt-wg1#21 ->(wq_completion)wg-kex-wg1#41 ->(wq_completion)wg-kex-wg1#42 ->(wq_completion)wg-crypt-wg0#21 ->(wq_completion)wg-kex-wg0#41 ->(wq_completion)wg-kex-wg0#42 ->(wq_completion)wg-crypt-wg2#22 ->(wq_completion)wg-kex-wg2#44 ->(wq_completion)wg-kex-wg2#43 ->(wq_completion)wg-crypt-wg1#22 ->(wq_completion)wg-kex-wg1#43 ->(wq_completion)wg-kex-wg1#44 ->(wq_completion)wg-crypt-wg0#22 ->(wq_completion)wg-kex-wg0#43 ->(wq_completion)wg-kex-wg0#44 ->(wq_completion)wg-crypt-wg2#23 ->(wq_completion)wg-kex-wg2#46 ->(wq_completion)wg-kex-wg2#45 ->(wq_completion)wg-crypt-wg1#23 ->(wq_completion)wg-kex-wg1#45 ->(wq_completion)wg-kex-wg1#46 ->(wq_completion)wg-crypt-wg0#23 ->(wq_completion)wg-kex-wg0#45 ->(wq_completion)wg-kex-wg0#46 ->(wq_completion)wg-crypt-wg2#24 ->(wq_completion)wg-kex-wg2#48 ->(wq_completion)wg-kex-wg2#47 ->(wq_completion)wg-crypt-wg1#24 ->(wq_completion)wg-kex-wg1#48 ->(wq_completion)wg-kex-wg1#47 ->(wq_completion)wg-crypt-wg0#24 ->(wq_completion)wg-kex-wg0#47 ->(wq_completion)wg-kex-wg0#48 ->batched_entropy_u8.lock ->kfence_freelist_lock ->(wq_completion)wg-crypt-wg2#16 ->(wq_completion)wg-kex-wg2#32 ->(wq_completion)wg-kex-wg2#31 ->(wq_completion)wg-crypt-wg1#16 ->(wq_completion)wg-kex-wg1#31 ->(wq_completion)wg-kex-wg1#32 ->(wq_completion)wg-crypt-wg0#16 ->(wq_completion)wg-kex-wg0#31 ->(wq_completion)wg-kex-wg0#32 ->(wq_completion)wg-crypt-wg2#26 ->(wq_completion)wg-kex-wg2#51 ->(wq_completion)wg-kex-wg2#52 ->(wq_completion)wg-crypt-wg1#26 ->(wq_completion)wg-kex-wg1#52 ->(wq_completion)wg-kex-wg1#51 ->(wq_completion)wg-crypt-wg0#26 ->(wq_completion)wg-kex-wg0#51 ->(wq_completion)wg-kex-wg0#52 ->(wq_completion)wg-crypt-wg2#25 ->(wq_completion)wg-kex-wg2#50 ->(wq_completion)wg-kex-wg2#49 ->rcu_state.barrier_mutex.wait_lock ->(wq_completion)wg-crypt-wg1#25 ->(wq_completion)wg-kex-wg1#50 ->(wq_completion)wg-kex-wg1#49 ->(wq_completion)wg-crypt-wg0#25 ->(wq_completion)wg-kex-wg0#49 ->(wq_completion)wg-kex-wg0#50 ->(wq_completion)wg-crypt-wg2#20 ->(wq_completion)wg-kex-wg2#40 ->(wq_completion)wg-kex-wg2#39 ->(wq_completion)wg-crypt-wg1#20 ->(wq_completion)wg-kex-wg1#39 ->(wq_completion)wg-kex-wg1#40 ->(wq_completion)wg-crypt-wg0#20 ->(wq_completion)wg-kex-wg0#39 ->(wq_completion)wg-kex-wg0#40 ->(wq_completion)wg-crypt-wg2#27 ->(wq_completion)wg-kex-wg2#53 ->(wq_completion)wg-kex-wg2#54 ->(wq_completion)wg-crypt-wg1#27 ->(wq_completion)wg-kex-wg1#53 ->(wq_completion)wg-kex-wg1#54 ->(wq_completion)wg-crypt-wg0#27 ->(wq_completion)wg-kex-wg0#53 ->(wq_completion)wg-kex-wg0#54 ->(wq_completion)wg-crypt-wg2#30 ->(wq_completion)wg-kex-wg2#60 ->(wq_completion)wg-kex-wg2#59 ->(wq_completion)wg-crypt-wg1#30 ->(wq_completion)wg-kex-wg1#59 ->(wq_completion)wg-kex-wg1#60 ->(wq_completion)wg-crypt-wg0#30 ->(wq_completion)wg-kex-wg0#59 ->(wq_completion)wg-kex-wg0#60 ->(wq_completion)wg-crypt-wg2#29 ->(wq_completion)wg-kex-wg2#57 ->(wq_completion)wg-kex-wg2#58 ->(wq_completion)wg-crypt-wg1#29 ->(wq_completion)wg-kex-wg1#57 ->(wq_completion)wg-kex-wg1#58 ->(wq_completion)wg-crypt-wg0#29 ->(wq_completion)wg-kex-wg0#57 ->(wq_completion)wg-kex-wg0#58 ->(wq_completion)wg-crypt-wg2#18 ->(wq_completion)wg-kex-wg2#36 ->(wq_completion)wg-kex-wg2#35 ->(wq_completion)wg-crypt-wg1#18 ->(wq_completion)wg-kex-wg1#35 ->(wq_completion)wg-kex-wg1#36 ->(wq_completion)wg-crypt-wg0#18 ->(wq_completion)wg-kex-wg0#35 ->(wq_completion)wg-kex-wg0#36 ->(wq_completion)wg-crypt-wg2#31 ->(wq_completion)wg-kex-wg2#62 ->(wq_completion)wg-kex-wg2#61 ->(wq_completion)wg-crypt-wg1#31 ->(wq_completion)wg-kex-wg1#61 ->(wq_completion)wg-kex-wg1#62 ->(wq_completion)wg-crypt-wg0#31 ->(wq_completion)wg-kex-wg0#61 ->(wq_completion)wg-kex-wg0#62 ->(wq_completion)wg-crypt-wg2#32 ->(wq_completion)wg-kex-wg2#63 ->(wq_completion)wg-kex-wg2#64 ->(wq_completion)wg-crypt-wg1#32 ->(wq_completion)wg-kex-wg1#63 ->(wq_completion)wg-kex-wg1#64 ->(wq_completion)wg-crypt-wg0#32 ->(wq_completion)wg-kex-wg0#63 ->(wq_completion)wg-kex-wg0#64 ->(wq_completion)wg-crypt-wg2#28 ->(wq_completion)wg-kex-wg2#55 ->(wq_completion)wg-kex-wg2#56 ->(wq_completion)wg-crypt-wg1#28 ->(wq_completion)wg-kex-wg1#55 ->(wq_completion)wg-kex-wg1#56 ->(wq_completion)wg-crypt-wg0#28 ->(wq_completion)wg-kex-wg0#55 ->(wq_completion)wg-kex-wg0#56 ->(wq_completion)wg-crypt-wg2#37 ->(wq_completion)wg-kex-wg2#73 ->(wq_completion)wg-kex-wg2#74 ->(wq_completion)wg-crypt-wg1#37 ->(wq_completion)wg-kex-wg1#73 ->(wq_completion)wg-kex-wg1#74 ->(wq_completion)wg-crypt-wg0#37 ->(wq_completion)wg-kex-wg0#73 ->(wq_completion)wg-kex-wg0#74 FD: 51 BD: 255 ++++: &wg->static_identity.lock ->&handshake->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&sem->wait_lock ->&p->pi_lock FD: 49 BD: 256 ++++: &handshake->lock ->crngs.lock ->tk_core.seq.seqcount ->&table->lock#2 ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->remove_cache_srcu ->&sem->wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 46 +.+.: &table->lock FD: 50 BD: 257 ++-.: &peer->endpoint_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 14 BD: 45 +.-.: (&app->periodic_timer) ->&app->lock#2 FD: 1 BD: 45 +...: _xmit_SIT FD: 16 BD: 53 +...: &bridge_netdev_addr_lock_key/1 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 39 BD: 45 +.-.: (&brmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 38 BD: 63 +.-.: &br->multicast_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&c->lock ->&dir->lock#2 ->deferred_lock ->&____s->seqcount ->nl_table_lock ->nl_table_wait.lock ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock FD: 39 BD: 45 +.-.: (&brmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 54 BD: 1 +.-.: (&in_dev->mr_ifc_timer) ->&obj_hash[i].lock ->batched_entropy_u32.lock ->&base->lock FD: 30 BD: 1 ..-.: &(&br->gc_work)->timer FD: 12 BD: 46 +.+.: (work_completion)(&(&br->gc_work)->work) ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 45 +...: _xmit_TUNNEL FD: 24 BD: 5 +.+.: (work_completion)(&tn->work) ->&rq->__lock FD: 16 BD: 45 +...: _xmit_IPGRE ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&c->lock ->&n->list_lock FD: 1 BD: 10 +.+.: nsim_bus_dev_list_lock.wait_lock FD: 1 BD: 45 +...: _xmit_TUNNEL6 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#335 FD: 30 BD: 1 ..-.: &(&hdev->cmd_timer)->timer FD: 41 BD: 48 +.+.: (work_completion)(&(&hdev->cmd_timer)->work) ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 43 BD: 53 +...: &dev_addr_list_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1460 FD: 1 BD: 246 ..-.: &list->lock#12 FD: 43 BD: 53 +...: &dev_addr_list_lock_key#2/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 39 BD: 1 +.-.: (&pmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 39 BD: 1 +.-.: (&pmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#22 ->&rq->__lock FD: 48 BD: 48 +.-.: _xmit_TUNNEL6#2 ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock FD: 30 BD: 1 ..-.: drivers/regulator/core.c:6335 FD: 4 BD: 2 +.+.: (regulator_init_complete_work).work ->&k->list_lock ->&k->k_lock FD: 43 BD: 53 +...: _xmit_ETHER/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->pool_lock#2 ->&n->list_lock FD: 19 BD: 55 +.-.: &hsr->seqnr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 56 +.-.: &new_node->seq_out_lock FD: 21 BD: 45 +.-.: (&hsr->announce_timer) FD: 1 BD: 45 +.+.: &nn->netlink_tap_lock FD: 16 BD: 53 +...: &batadv_netdev_addr_lock_key/1 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 43 BD: 53 +...: &vlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 43 BD: 53 +...: &macvlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 47 +.-.: &list->lock#13 FD: 29 BD: 46 +.+.: (work_completion)(&port->bc_work) ->&list->lock#13 ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->&base->lock FD: 25 BD: 92 -.-.: &rq_wait->wait ->&p->pi_lock FD: 16 BD: 46 +...: &ipvlan->addrs_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 43 BD: 53 +...: &macsec_netdev_addr_lock_key/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 16 BD: 56 +.-.: key#19 ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 23 BD: 49 +...: &bat_priv->tt.commit_lock ->key#15 ->&bat_priv->softif_vlan_list_lock ->&bat_priv->tt.changes_list_lock ->&bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 13 BD: 45 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&sch->q.lock FD: 1 BD: 46 +.+.: &wg->socket_update_lock FD: 1 BD: 188 +.-.: &list->lock#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1596 FD: 51 BD: 115 +.+.: (work_completion)(&peer->transmit_handshake_work) ->tk_core.seq.seqcount ->&wg->static_identity.lock ->&cookie->lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->pool_lock#2 ->&peer->endpoint_lock ->batched_entropy_u8.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->kfence_freelist_lock ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 258 +...: &table->lock#2 FD: 24 BD: 255 ++++: &cookie->lock ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1593 ->&rq->__lock FD: 1 BD: 188 +.-.: &r->producer_lock#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1444 ->&rq->__lock FD: 84 BD: 184 +.+.: (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&r->consumer_lock#2 ->&wg->static_identity.lock ->&peer->endpoint_lock ->tk_core.seq.seqcount ->&cookie->lock ->&handshake->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&list->lock#14 ->&rq->__lock ->rcu_node_0 ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&cfs_rq->removed.lock ->&n->list_lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->quarantine_lock FD: 1 BD: 185 +.+.: &r->consumer_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1589 FD: 5 BD: 257 +.-.: &peer->keypairs.keypair_update_lock ->&table->lock#2 ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1591 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1592 ->&rq->__lock FD: 25 BD: 115 +.+.: (work_completion)(&peer->transmit_packet_work) ->&obj_hash[i].lock ->&peer->endpoint_lock ->&base->lock ->batched_entropy_u8.lock ->&rq->__lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1591 FD: 1 BD: 1 +.-.: &keypair->receiving_counter.lock FD: 1 BD: 14 +.+.: &data->mtx FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1596 FD: 25 BD: 1 ++++: &iopt->domains_rwsem ->&iopt->iova_rwsem ->&rq->__lock FD: 1 BD: 56 +.-.: &entry->crc_lock FD: 2 BD: 1 +.-.: (&tun->flow_gc_timer) ->&tun->lock FD: 1 BD: 46 +.-.: &tun->lock FD: 1 BD: 49 ....: &wdev->event_lock FD: 1 BD: 49 ....: (&dwork->timer) FD: 1 BD: 49 ....: (&dwork->timer)#2 FD: 1 BD: 49 +.+.: (work_completion)(&(&link->color_collision_detect_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1976 FD: 1 BD: 50 ..-.: &list->lock#15 FD: 1 BD: 49 +.-.: &ifibss->incomplete_lock FD: 852 BD: 1 +.+.: (wq_completion)cfg80211 ->(work_completion)(&rdev->event_work) ->(work_completion)(&(&rdev->dfs_update_channels_wk)->work) FD: 400 BD: 6 +.+.: (work_completion)(&rdev->event_work) ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock FD: 36 BD: 2 +.+.: wireless_nlevent_work ->net_rwsem FD: 80 BD: 1 +.+.: &type->s_umount_key#46/1 ->fs_reclaim ->&c->lock ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#32 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->binderfs_minors_mutex ->&dentry->d_lock ->&____s->seqcount ->&sb->s_type->i_mutex_key#18 ->&n->list_lock ->&____s->seqcount#2 ->&rq->__lock FD: 37 BD: 4 +.+.: &sb->s_type->i_lock_key#32 ->&dentry->d_lock FD: 25 BD: 3 +.+.: binderfs_minors_mutex ->binderfs_minors.xa_lock ->&rq->__lock FD: 1 BD: 4 ....: binderfs_minors.xa_lock FD: 64 BD: 2 +.+.: &sb->s_type->i_mutex_key#18 ->&sb->s_type->i_lock_key#32 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->&rq->__lock FD: 1 BD: 3 +.+.: iunique_lock FD: 789 BD: 2 +.+.: &type->i_mutex_dir_key#6/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex ->&c->lock FD: 51 BD: 1 .+.+: kn->active#48 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 49 BD: 1 ++++: kn->active#49 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu FD: 3 BD: 89 ..-.: cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&obj_hash[i].lock FD: 9 BD: 50 +...: &bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1426 FD: 1 BD: 1 +...: bpf_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1425 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1422 FD: 3 BD: 89 ....: kernfs_pr_cont_lock ->kernfs_rename_lock ->(console_sem).lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#530 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 73 +.-.: &local->active_txq_lock[i] FD: 38 BD: 72 +.-.: &local->handle_wake_tx_queue_lock ->&local->active_txq_lock[i] ->&local->queue_stop_reason_lock ->&fq->lock ->tk_core.seq.seqcount ->hwsim_radio_lock ->&list->lock#16 FD: 1 BD: 73 ..-.: &local->queue_stop_reason_lock FD: 1 BD: 75 ..-.: &list->lock#16 FD: 32 BD: 1 +.-.: &local->rx_path_lock ->&obj_hash[i].lock ->pool_lock#2 ->&list->lock#15 ->&rdev->wiphy_work_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1941 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#74 FD: 16 BD: 49 +...: &sta->lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 16 BD: 49 +...: &sta->rate_ctrl_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 25 BD: 89 ....: cgroup_threadgroup_rwsem.waiters.lock ->&p->pi_lock FD: 1 BD: 16 +.+.: (wq_completion)cpuset_migrate_mm FD: 789 BD: 2 +.+.: &type->i_mutex_dir_key#7/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex ->&c->lock ->pool_lock#2 ->&xa->xa_lock#5 ->&obj_hash[i].lock ->stock_lock FD: 49 BD: 1 ++++: kn->active#50 ->fs_reclaim ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&kernfs_locks->open_file_mutex[count] ->stock_lock ->&n->list_lock ->remove_cache_srcu FD: 1 BD: 90 ....: cpuset_attach_wq.lock FD: 2 BD: 567 ..-.: stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 51 BD: 1 .+.+: kn->active#51 ->fs_reclaim ->stock_lock ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#52 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->memcg_max_mutex ->&c->lock ->&n->list_lock FD: 1 BD: 8 +.+.: memcg_max_mutex FD: 1 BD: 6 ....: &per_cpu(xt_recseq, i) FD: 229 BD: 5 +.+.: nf_nat_proto_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->stock_lock ->&rq->__lock ->&n->list_lock ->krc.lock FD: 26 BD: 1 +.+.: loop_validate_mutex ->&lo->lo_mutex ->&rq->__lock ->loop_validate_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 57 +.-.: &nf_nat_locks[i] FD: 13 BD: 45 +.-.: (&hsr->prune_timer) ->&hsr->list_lock ->&obj_hash[i].lock ->&base->lock FD: 1778 BD: 1 +.+.: &ndev->req_lock ->&wq->mutex ->(&ndev->cmd_timer) ->&obj_hash[i].lock ->&base->lock ->(&ndev->data_timer) ->pool_lock#2 ->&rq->__lock ->&cfs_rq->removed.lock ->(wq_completion)nfc3_nci_cmd_wq#1220 ->(wq_completion)nfc2_nci_cmd_wq#1361 ->(wq_completion)nfc3_nci_cmd_wq#1221 ->(wq_completion)nfc2_nci_cmd_wq#1362 ->(wq_completion)nfc3_nci_cmd_wq#1222 ->(wq_completion)nfc2_nci_cmd_wq#1363 ->(wq_completion)nfc3_nci_cmd_wq#1223 ->(wq_completion)nfc2_nci_cmd_wq#1364 ->(wq_completion)nfc3_nci_cmd_wq#1224 ->(wq_completion)nfc2_nci_cmd_wq#1365 ->(wq_completion)nfc3_nci_cmd_wq#1225 ->(wq_completion)nfc2_nci_cmd_wq#1366 ->(wq_completion)nfc3_nci_cmd_wq#1226 ->(wq_completion)nfc2_nci_cmd_wq#1367 ->(wq_completion)nfc3_nci_cmd_wq#1227 ->(wq_completion)nfc2_nci_cmd_wq#1368 ->(wq_completion)nfc3_nci_cmd_wq#1228 ->(wq_completion)nfc2_nci_cmd_wq#1369 ->(wq_completion)nfc3_nci_cmd_wq#1229 ->(wq_completion)nfc2_nci_cmd_wq#1370 ->(wq_completion)nfc3_nci_cmd_wq#1230 ->(wq_completion)nfc2_nci_cmd_wq#1371 ->(wq_completion)nfc3_nci_cmd_wq#1231 ->(wq_completion)nfc2_nci_cmd_wq#1372 ->(wq_completion)nfc3_nci_cmd_wq#1232 ->(wq_completion)nfc4_nci_cmd_wq#346 ->(wq_completion)nfc5_nci_cmd_wq#247 ->(wq_completion)nfc2_nci_cmd_wq#1373 ->(wq_completion)nfc3_nci_cmd_wq#1233 ->(wq_completion)nfc2_nci_cmd_wq#1374 ->(wq_completion)nfc3_nci_cmd_wq#1234 ->(wq_completion)nfc2_nci_cmd_wq#1375 ->(wq_completion)nfc3_nci_cmd_wq#1235 ->(wq_completion)nfc4_nci_cmd_wq#347 ->(wq_completion)nfc5_nci_cmd_wq#248 ->(wq_completion)nfc2_nci_cmd_wq#1376 ->(wq_completion)nfc3_nci_cmd_wq#1236 ->(wq_completion)nfc2_nci_cmd_wq#1377 ->(wq_completion)nfc3_nci_cmd_wq#1237 ->(wq_completion)nfc4_nci_cmd_wq#348 ->(wq_completion)nfc5_nci_cmd_wq#249 ->(wq_completion)nfc2_nci_cmd_wq#1378 ->(wq_completion)nfc3_nci_cmd_wq#1238 ->(wq_completion)nfc2_nci_cmd_wq#1379 ->(wq_completion)nfc3_nci_cmd_wq#1239 ->(wq_completion)nfc2_nci_cmd_wq#1380 ->(wq_completion)nfc3_nci_cmd_wq#1240 ->(wq_completion)nfc2_nci_cmd_wq#1381 ->(wq_completion)nfc3_nci_cmd_wq#1241 ->(wq_completion)nfc2_nci_cmd_wq#1382 ->(wq_completion)nfc3_nci_cmd_wq#1242 ->(wq_completion)nfc2_nci_cmd_wq#1383 ->(wq_completion)nfc3_nci_cmd_wq#1243 ->(wq_completion)nfc2_nci_cmd_wq#1384 ->(wq_completion)nfc3_nci_cmd_wq#1244 ->(wq_completion)nfc2_nci_cmd_wq#1385 ->(wq_completion)nfc3_nci_cmd_wq#1245 ->(wq_completion)nfc2_nci_cmd_wq#1386 ->(wq_completion)nfc3_nci_cmd_wq#1246 ->(wq_completion)nfc2_nci_cmd_wq#1387 ->(wq_completion)nfc3_nci_cmd_wq#1247 ->(wq_completion)nfc4_nci_cmd_wq#349 ->(wq_completion)nfc5_nci_cmd_wq#250 ->(wq_completion)nfc2_nci_cmd_wq#1388 ->(wq_completion)nfc3_nci_cmd_wq#1248 ->(wq_completion)nfc2_nci_cmd_wq#1389 ->(wq_completion)nfc4_nci_cmd_wq#350 ->(wq_completion)nfc3_nci_cmd_wq#1249 ->(wq_completion)nfc5_nci_cmd_wq#251 ->(wq_completion)nfc2_nci_cmd_wq#1390 ->(wq_completion)nfc3_nci_cmd_wq#1250 ->(wq_completion)nfc2_nci_cmd_wq#1391 ->(wq_completion)nfc3_nci_cmd_wq#1251 ->(wq_completion)nfc2_nci_cmd_wq#1392 ->(wq_completion)nfc3_nci_cmd_wq#1252 ->(wq_completion)nfc4_nci_cmd_wq#351 ->(wq_completion)nfc2_nci_cmd_wq#1393 ->(wq_completion)nfc2_nci_cmd_wq#1394 ->(wq_completion)nfc3_nci_cmd_wq#1253 ->(wq_completion)nfc2_nci_cmd_wq#1395 ->(wq_completion)nfc3_nci_cmd_wq#1254 ->(wq_completion)nfc4_nci_cmd_wq#352 ->(wq_completion)nfc5_nci_cmd_wq#252 ->(wq_completion)nfc2_nci_cmd_wq#1396 ->(wq_completion)nfc3_nci_cmd_wq#1255 ->(wq_completion)nfc2_nci_cmd_wq#1397 ->(wq_completion)nfc3_nci_cmd_wq#1256 ->(wq_completion)nfc2_nci_cmd_wq#1398 ->(wq_completion)nfc3_nci_cmd_wq#1257 ->(wq_completion)nfc2_nci_cmd_wq#1399 ->(wq_completion)nfc3_nci_cmd_wq#1258 ->(wq_completion)nfc2_nci_cmd_wq#1400 ->(wq_completion)nfc3_nci_cmd_wq#1259 ->(wq_completion)nfc2_nci_cmd_wq#1401 ->(wq_completion)nfc3_nci_cmd_wq#1260 ->(wq_completion)nfc2_nci_cmd_wq#1402 ->(wq_completion)nfc3_nci_cmd_wq#1261 ->(wq_completion)nfc2_nci_cmd_wq#1403 ->(wq_completion)nfc3_nci_cmd_wq#1262 ->(wq_completion)nfc2_nci_cmd_wq#1404 ->(wq_completion)nfc3_nci_cmd_wq#1263 ->(wq_completion)nfc3_nci_cmd_wq#1264 ->(wq_completion)nfc4_nci_cmd_wq#353 ->(wq_completion)nfc2_nci_cmd_wq#1405 ->(wq_completion)nfc5_nci_cmd_wq#253 ->(wq_completion)nfc2_nci_cmd_wq#1406 ->(wq_completion)nfc3_nci_cmd_wq#1265 ->(wq_completion)nfc2_nci_cmd_wq#1407 ->(wq_completion)nfc3_nci_cmd_wq#1266 ->(wq_completion)nfc4_nci_cmd_wq#354 ->(wq_completion)nfc2_nci_cmd_wq#1408 ->(wq_completion)nfc2_nci_cmd_wq#1409 ->(wq_completion)nfc3_nci_cmd_wq#1267 ->(wq_completion)nfc2_nci_cmd_wq#1410 ->(wq_completion)nfc3_nci_cmd_wq#1268 ->(wq_completion)nfc2_nci_cmd_wq#1411 ->(wq_completion)nfc3_nci_cmd_wq#1269 ->(wq_completion)nfc2_nci_cmd_wq#1412 ->(wq_completion)nfc3_nci_cmd_wq#1270 ->(wq_completion)nfc2_nci_cmd_wq#1413 ->(wq_completion)nfc3_nci_cmd_wq#1271 ->(wq_completion)nfc2_nci_cmd_wq#1414 ->(wq_completion)nfc3_nci_cmd_wq#1272 ->(wq_completion)nfc2_nci_cmd_wq#1415 ->(wq_completion)nfc3_nci_cmd_wq#1273 ->(wq_completion)nfc2_nci_cmd_wq#1416 ->(wq_completion)nfc3_nci_cmd_wq#1274 ->(wq_completion)nfc2_nci_cmd_wq#1417 ->(wq_completion)nfc3_nci_cmd_wq#1275 ->(wq_completion)nfc2_nci_cmd_wq#1418 ->(wq_completion)nfc3_nci_cmd_wq#1276 ->(wq_completion)nfc2_nci_cmd_wq#1419 ->(wq_completion)nfc3_nci_cmd_wq#1277 ->(wq_completion)nfc2_nci_cmd_wq#1420 ->(wq_completion)nfc3_nci_cmd_wq#1278 ->(wq_completion)nfc2_nci_cmd_wq#1421 ->(wq_completion)nfc3_nci_cmd_wq#1279 ->(wq_completion)nfc2_nci_cmd_wq#1422 ->(wq_completion)nfc3_nci_cmd_wq#1280 ->(wq_completion)nfc2_nci_cmd_wq#1423 ->(wq_completion)nfc3_nci_cmd_wq#1281 ->(wq_completion)nfc2_nci_cmd_wq#1424 ->(wq_completion)nfc3_nci_cmd_wq#1282 ->(wq_completion)nfc2_nci_cmd_wq#1425 ->(wq_completion)nfc3_nci_cmd_wq#1283 ->(wq_completion)nfc2_nci_cmd_wq#1426 ->(wq_completion)nfc3_nci_cmd_wq#1284 ->(wq_completion)nfc2_nci_cmd_wq#1427 ->(wq_completion)nfc3_nci_cmd_wq#1285 ->(wq_completion)nfc2_nci_cmd_wq#1428 ->(wq_completion)nfc4_nci_cmd_wq#355 ->(wq_completion)nfc3_nci_cmd_wq#1286 ->(wq_completion)nfc5_nci_cmd_wq#254 ->(wq_completion)nfc2_nci_cmd_wq#1429 ->(wq_completion)nfc3_nci_cmd_wq#1287 ->(wq_completion)nfc2_nci_cmd_wq#1430 ->(wq_completion)nfc3_nci_cmd_wq#1288 ->(wq_completion)nfc2_nci_cmd_wq#1431 ->(wq_completion)nfc3_nci_cmd_wq#1289 ->(wq_completion)nfc2_nci_cmd_wq#1432 ->(wq_completion)nfc3_nci_cmd_wq#1290 ->(wq_completion)nfc2_nci_cmd_wq#1433 ->(wq_completion)nfc3_nci_cmd_wq#1291 ->(wq_completion)nfc4_nci_cmd_wq#356 ->(wq_completion)nfc2_nci_cmd_wq#1434 ->(wq_completion)nfc2_nci_cmd_wq#1435 ->(wq_completion)nfc3_nci_cmd_wq#1292 ->(wq_completion)nfc2_nci_cmd_wq#1436 ->(wq_completion)nfc3_nci_cmd_wq#1293 ->(wq_completion)nfc2_nci_cmd_wq#1437 ->(wq_completion)nfc3_nci_cmd_wq#1294 ->(wq_completion)nfc4_nci_cmd_wq#357 ->(wq_completion)nfc3_nci_cmd_wq#1295 ->(wq_completion)nfc2_nci_cmd_wq#1438 ->(wq_completion)nfc5_nci_cmd_wq#255 ->(wq_completion)nfc2_nci_cmd_wq#1439 ->(wq_completion)nfc3_nci_cmd_wq#1296 ->(wq_completion)nfc2_nci_cmd_wq#1440 ->(wq_completion)nfc3_nci_cmd_wq#1297 ->(wq_completion)nfc2_nci_cmd_wq#1441 ->(wq_completion)nfc3_nci_cmd_wq#1298 ->(wq_completion)nfc2_nci_cmd_wq#1442 ->(wq_completion)nfc3_nci_cmd_wq#1299 ->(wq_completion)nfc2_nci_cmd_wq#1443 ->(wq_completion)nfc3_nci_cmd_wq#1300 ->(wq_completion)nfc2_nci_cmd_wq#1444 ->(wq_completion)nfc3_nci_cmd_wq#1301 ->(wq_completion)nfc2_nci_cmd_wq#1445 ->(wq_completion)nfc3_nci_cmd_wq#1302 ->(wq_completion)nfc2_nci_cmd_wq#1446 ->(wq_completion)nfc3_nci_cmd_wq#1303 ->(wq_completion)nfc2_nci_cmd_wq#1447 ->(wq_completion)nfc3_nci_cmd_wq#1304 ->(wq_completion)nfc2_nci_cmd_wq#1448 ->(wq_completion)nfc3_nci_cmd_wq#1305 ->(wq_completion)nfc2_nci_cmd_wq#1449 ->(wq_completion)nfc3_nci_cmd_wq#1306 ->(wq_completion)nfc2_nci_cmd_wq#1450 ->(wq_completion)nfc3_nci_cmd_wq#1307 ->(wq_completion)nfc5_nci_cmd_wq#256 ->(wq_completion)nfc4_nci_cmd_wq#358 ->(wq_completion)nfc2_nci_cmd_wq#1451 ->(wq_completion)nfc3_nci_cmd_wq#1308 ->(wq_completion)nfc2_nci_cmd_wq#1452 ->(wq_completion)nfc3_nci_cmd_wq#1309 ->(wq_completion)nfc2_nci_cmd_wq#1453 ->(wq_completion)nfc3_nci_cmd_wq#1310 ->(wq_completion)nfc2_nci_cmd_wq#1454 ->(wq_completion)nfc3_nci_cmd_wq#1311 ->(wq_completion)nfc2_nci_cmd_wq#1455 ->(wq_completion)nfc3_nci_cmd_wq#1312 ->(wq_completion)nfc2_nci_cmd_wq#1456 ->(wq_completion)nfc3_nci_cmd_wq#1313 ->(wq_completion)nfc2_nci_cmd_wq#1457 ->(wq_completion)nfc3_nci_cmd_wq#1314 ->(wq_completion)nfc2_nci_cmd_wq#1458 ->(wq_completion)nfc3_nci_cmd_wq#1315 ->(wq_completion)nfc2_nci_cmd_wq#1459 ->(wq_completion)nfc3_nci_cmd_wq#1316 ->(wq_completion)nfc2_nci_cmd_wq#1460 ->(wq_completion)nfc3_nci_cmd_wq#1317 ->(wq_completion)nfc2_nci_cmd_wq#1461 ->(wq_completion)nfc3_nci_cmd_wq#1318 ->(wq_completion)nfc2_nci_cmd_wq#1462 ->(wq_completion)nfc3_nci_cmd_wq#1319 ->(wq_completion)nfc2_nci_cmd_wq#1463 ->(wq_completion)nfc3_nci_cmd_wq#1320 ->(wq_completion)nfc2_nci_cmd_wq#1464 ->(wq_completion)nfc3_nci_cmd_wq#1321 ->(wq_completion)nfc2_nci_cmd_wq#1465 ->(wq_completion)nfc3_nci_cmd_wq#1322 ->(wq_completion)nfc2_nci_cmd_wq#1466 ->(wq_completion)nfc3_nci_cmd_wq#1323 ->(wq_completion)nfc2_nci_cmd_wq#1467 ->(wq_completion)nfc3_nci_cmd_wq#1324 ->(wq_completion)nfc2_nci_cmd_wq#1468 ->(wq_completion)nfc3_nci_cmd_wq#1325 ->(wq_completion)nfc2_nci_cmd_wq#1469 ->(wq_completion)nfc3_nci_cmd_wq#1326 ->(wq_completion)nfc2_nci_cmd_wq#1470 ->(wq_completion)nfc3_nci_cmd_wq#1327 ->(wq_completion)nfc2_nci_cmd_wq#1471 ->(wq_completion)nfc3_nci_cmd_wq#1328 ->(wq_completion)nfc2_nci_cmd_wq#1472 ->(wq_completion)nfc3_nci_cmd_wq#1329 ->(wq_completion)nfc2_nci_cmd_wq#1473 ->(wq_completion)nfc3_nci_cmd_wq#1330 ->(wq_completion)nfc2_nci_cmd_wq#1474 ->(wq_completion)nfc3_nci_cmd_wq#1331 ->(wq_completion)nfc4_nci_cmd_wq#359 ->(wq_completion)nfc5_nci_cmd_wq#257 ->(wq_completion)nfc2_nci_cmd_wq#1475 ->(wq_completion)nfc3_nci_cmd_wq#1332 ->(wq_completion)nfc2_nci_cmd_wq#1476 ->(wq_completion)nfc3_nci_cmd_wq#1333 ->(wq_completion)nfc2_nci_cmd_wq#1477 ->(wq_completion)nfc3_nci_cmd_wq#1334 ->(wq_completion)nfc2_nci_cmd_wq#1478 ->(wq_completion)nfc3_nci_cmd_wq#1335 ->(wq_completion)nfc2_nci_cmd_wq#1479 ->(wq_completion)nfc3_nci_cmd_wq#1336 ->(wq_completion)nfc2_nci_cmd_wq#1480 ->(wq_completion)nfc3_nci_cmd_wq#1337 ->(wq_completion)nfc2_nci_cmd_wq#1481 ->(wq_completion)nfc3_nci_cmd_wq#1338 ->(wq_completion)nfc2_nci_cmd_wq#1482 ->(wq_completion)nfc3_nci_cmd_wq#1339 ->(wq_completion)nfc2_nci_cmd_wq#1483 ->(wq_completion)nfc3_nci_cmd_wq#1340 ->(wq_completion)nfc2_nci_cmd_wq#1484 ->(wq_completion)nfc3_nci_cmd_wq#1341 ->(wq_completion)nfc4_nci_cmd_wq#360 ->(wq_completion)nfc5_nci_cmd_wq#258 ->(wq_completion)nfc6_nci_cmd_wq#46 ->(wq_completion)nfc7_nci_cmd_wq#35 ->(wq_completion)nfc2_nci_cmd_wq#1485 ->(wq_completion)nfc3_nci_cmd_wq#1342 ->(wq_completion)nfc2_nci_cmd_wq#1486 ->(wq_completion)nfc3_nci_cmd_wq#1343 ->(wq_completion)nfc2_nci_cmd_wq#1487 ->(wq_completion)nfc3_nci_cmd_wq#1344 ->(wq_completion)nfc2_nci_cmd_wq#1488 ->(wq_completion)nfc3_nci_cmd_wq#1345 ->(wq_completion)nfc2_nci_cmd_wq#1489 ->(wq_completion)nfc3_nci_cmd_wq#1346 ->(wq_completion)nfc2_nci_cmd_wq#1490 ->(wq_completion)nfc3_nci_cmd_wq#1347 ->(wq_completion)nfc2_nci_cmd_wq#1491 ->(wq_completion)nfc3_nci_cmd_wq#1348 ->(wq_completion)nfc2_nci_cmd_wq#1492 ->(wq_completion)nfc3_nci_cmd_wq#1349 ->(wq_completion)nfc2_nci_cmd_wq#1493 ->(wq_completion)nfc3_nci_cmd_wq#1350 ->(wq_completion)nfc2_nci_cmd_wq#1494 ->(wq_completion)nfc3_nci_cmd_wq#1351 ->(wq_completion)nfc2_nci_cmd_wq#1495 ->(wq_completion)nfc3_nci_cmd_wq#1352 ->(wq_completion)nfc2_nci_cmd_wq#1496 ->(wq_completion)nfc3_nci_cmd_wq#1353 ->(wq_completion)nfc4_nci_cmd_wq#361 ->(wq_completion)nfc2_nci_cmd_wq#1497 ->(wq_completion)nfc3_nci_cmd_wq#1354 ->(wq_completion)nfc2_nci_cmd_wq#1498 ->(wq_completion)nfc2_nci_cmd_wq#1499 ->(wq_completion)nfc3_nci_cmd_wq#1355 ->(wq_completion)nfc2_nci_cmd_wq#1500 ->(wq_completion)nfc3_nci_cmd_wq#1356 ->(wq_completion)nfc2_nci_cmd_wq#1501 ->(wq_completion)nfc3_nci_cmd_wq#1357 ->(wq_completion)nfc2_nci_cmd_wq#1502 ->(wq_completion)nfc3_nci_cmd_wq#1358 ->(wq_completion)nfc2_nci_cmd_wq#1503 ->(wq_completion)nfc3_nci_cmd_wq#1359 ->(wq_completion)nfc2_nci_cmd_wq#1504 ->(wq_completion)nfc3_nci_cmd_wq#1360 ->(wq_completion)nfc2_nci_cmd_wq#1505 ->(wq_completion)nfc3_nci_cmd_wq#1361 ->(wq_completion)nfc3_nci_cmd_wq#1362 ->(wq_completion)nfc2_nci_cmd_wq#1506 ->(wq_completion)nfc4_nci_cmd_wq#362 ->(wq_completion)nfc5_nci_cmd_wq#259 ->(wq_completion)nfc2_nci_cmd_wq#1507 ->(wq_completion)nfc3_nci_cmd_wq#1363 ->(wq_completion)nfc2_nci_cmd_wq#1508 ->(wq_completion)nfc3_nci_cmd_wq#1364 ->(wq_completion)nfc2_nci_cmd_wq#1509 ->(wq_completion)nfc3_nci_cmd_wq#1365 ->(wq_completion)nfc2_nci_cmd_wq#1510 ->(wq_completion)nfc3_nci_cmd_wq#1366 ->(wq_completion)nfc2_nci_cmd_wq#1511 ->(wq_completion)nfc3_nci_cmd_wq#1367 ->(wq_completion)nfc2_nci_cmd_wq#1512 ->(wq_completion)nfc3_nci_cmd_wq#1368 ->(wq_completion)nfc2_nci_cmd_wq#1513 ->(wq_completion)nfc3_nci_cmd_wq#1369 ->(wq_completion)nfc2_nci_cmd_wq#1514 ->(wq_completion)nfc3_nci_cmd_wq#1370 ->(wq_completion)nfc4_nci_cmd_wq#363 ->(wq_completion)nfc2_nci_cmd_wq#1515 ->(wq_completion)nfc2_nci_cmd_wq#1516 ->(wq_completion)nfc3_nci_cmd_wq#1371 ->(wq_completion)nfc4_nci_cmd_wq#364 ->(wq_completion)nfc2_nci_cmd_wq#1517 ->(wq_completion)nfc2_nci_cmd_wq#1518 ->(wq_completion)nfc3_nci_cmd_wq#1372 ->(wq_completion)nfc2_nci_cmd_wq#1519 ->(wq_completion)nfc3_nci_cmd_wq#1373 ->(wq_completion)nfc2_nci_cmd_wq#1520 ->(wq_completion)nfc3_nci_cmd_wq#1374 ->(wq_completion)nfc2_nci_cmd_wq#1521 ->(wq_completion)nfc3_nci_cmd_wq#1375 ->(wq_completion)nfc2_nci_cmd_wq#1522 ->(wq_completion)nfc3_nci_cmd_wq#1376 ->(wq_completion)nfc2_nci_cmd_wq#1523 ->(wq_completion)nfc3_nci_cmd_wq#1377 ->(wq_completion)nfc3_nci_cmd_wq#1378 ->(wq_completion)nfc4_nci_cmd_wq#365 ->(wq_completion)nfc2_nci_cmd_wq#1524 ->(wq_completion)nfc5_nci_cmd_wq#260 ->(wq_completion)nfc2_nci_cmd_wq#1525 ->(wq_completion)nfc3_nci_cmd_wq#1379 ->(wq_completion)nfc2_nci_cmd_wq#1526 ->(wq_completion)nfc3_nci_cmd_wq#1380 ->(wq_completion)nfc4_nci_cmd_wq#366 ->(wq_completion)nfc2_nci_cmd_wq#1527 ->(wq_completion)nfc2_nci_cmd_wq#1528 ->(wq_completion)nfc3_nci_cmd_wq#1381 ->(wq_completion)nfc2_nci_cmd_wq#1529 ->(wq_completion)nfc3_nci_cmd_wq#1382 ->(wq_completion)nfc2_nci_cmd_wq#1530 ->(wq_completion)nfc3_nci_cmd_wq#1383 ->(wq_completion)nfc4_nci_cmd_wq#367 ->(wq_completion)nfc2_nci_cmd_wq#1531 ->(wq_completion)nfc2_nci_cmd_wq#1532 ->(wq_completion)nfc3_nci_cmd_wq#1384 ->(wq_completion)nfc2_nci_cmd_wq#1533 ->(wq_completion)nfc3_nci_cmd_wq#1385 ->(wq_completion)nfc4_nci_cmd_wq#368 ->(wq_completion)nfc5_nci_cmd_wq#261 ->(wq_completion)nfc2_nci_cmd_wq#1534 ->(wq_completion)nfc3_nci_cmd_wq#1386 ->(wq_completion)nfc2_nci_cmd_wq#1535 ->(wq_completion)nfc3_nci_cmd_wq#1387 ->(wq_completion)nfc2_nci_cmd_wq#1536 ->(wq_completion)nfc3_nci_cmd_wq#1388 ->(wq_completion)nfc2_nci_cmd_wq#1537 ->(wq_completion)nfc3_nci_cmd_wq#1389 ->(wq_completion)nfc2_nci_cmd_wq#1538 ->(wq_completion)nfc3_nci_cmd_wq#1390 ->(wq_completion)nfc2_nci_cmd_wq#1539 ->(wq_completion)nfc3_nci_cmd_wq#1391 ->(wq_completion)nfc2_nci_cmd_wq#1540 ->(wq_completion)nfc3_nci_cmd_wq#1392 ->(wq_completion)nfc2_nci_cmd_wq#1541 ->(wq_completion)nfc3_nci_cmd_wq#1393 ->(wq_completion)nfc2_nci_cmd_wq#1542 ->(wq_completion)nfc3_nci_cmd_wq#1394 ->(wq_completion)nfc2_nci_cmd_wq#1543 ->(wq_completion)nfc3_nci_cmd_wq#1395 ->(wq_completion)nfc2_nci_cmd_wq#1544 ->(wq_completion)nfc3_nci_cmd_wq#1396 ->(wq_completion)nfc2_nci_cmd_wq#1545 ->(wq_completion)nfc3_nci_cmd_wq#1397 ->(wq_completion)nfc2_nci_cmd_wq#1546 ->(wq_completion)nfc3_nci_cmd_wq#1398 ->(wq_completion)nfc2_nci_cmd_wq#1547 ->(wq_completion)nfc3_nci_cmd_wq#1399 ->(wq_completion)nfc2_nci_cmd_wq#1548 ->(wq_completion)nfc3_nci_cmd_wq#1400 ->(wq_completion)nfc2_nci_cmd_wq#1549 ->(wq_completion)nfc3_nci_cmd_wq#1401 ->(wq_completion)nfc2_nci_cmd_wq#1550 ->(wq_completion)nfc3_nci_cmd_wq#1402 ->(wq_completion)nfc2_nci_cmd_wq#1551 ->(wq_completion)nfc3_nci_cmd_wq#1403 ->(wq_completion)nfc2_nci_cmd_wq#1552 ->(wq_completion)nfc3_nci_cmd_wq#1404 ->(wq_completion)nfc2_nci_cmd_wq#1553 ->(wq_completion)nfc3_nci_cmd_wq#1405 ->(wq_completion)nfc2_nci_cmd_wq#1554 ->(wq_completion)nfc3_nci_cmd_wq#1406 ->(wq_completion)nfc2_nci_cmd_wq#1555 ->(wq_completion)nfc3_nci_cmd_wq#1407 ->(wq_completion)nfc2_nci_cmd_wq#1556 ->(wq_completion)nfc3_nci_cmd_wq#1408 ->(wq_completion)nfc2_nci_cmd_wq#1557 ->(wq_completion)nfc3_nci_cmd_wq#1409 ->(wq_completion)nfc2_nci_cmd_wq#1558 ->(wq_completion)nfc3_nci_cmd_wq#1410 ->(wq_completion)nfc2_nci_cmd_wq#1559 ->(wq_completion)nfc3_nci_cmd_wq#1411 ->(wq_completion)nfc4_nci_cmd_wq#369 ->(wq_completion)nfc2_nci_cmd_wq#1560 ->(wq_completion)nfc2_nci_cmd_wq#1561 ->(wq_completion)nfc3_nci_cmd_wq#1412 ->(wq_completion)nfc2_nci_cmd_wq#1562 ->(wq_completion)nfc3_nci_cmd_wq#1413 ->(wq_completion)nfc2_nci_cmd_wq#1563 ->(wq_completion)nfc3_nci_cmd_wq#1414 ->(wq_completion)nfc2_nci_cmd_wq#1564 ->(wq_completion)nfc3_nci_cmd_wq#1415 ->(wq_completion)nfc2_nci_cmd_wq#1565 ->(wq_completion)nfc3_nci_cmd_wq#1416 ->(wq_completion)nfc2_nci_cmd_wq#1566 ->(wq_completion)nfc3_nci_cmd_wq#1417 ->(wq_completion)nfc2_nci_cmd_wq#1567 ->(wq_completion)nfc3_nci_cmd_wq#1418 ->(wq_completion)nfc2_nci_cmd_wq#1568 ->(wq_completion)nfc3_nci_cmd_wq#1419 ->(wq_completion)nfc2_nci_cmd_wq#1569 ->(wq_completion)nfc3_nci_cmd_wq#1420 ->(wq_completion)nfc2_nci_cmd_wq#1570 ->(wq_completion)nfc3_nci_cmd_wq#1421 ->(wq_completion)nfc2_nci_cmd_wq#1571 ->(wq_completion)nfc3_nci_cmd_wq#1422 ->(wq_completion)nfc2_nci_cmd_wq#1572 ->(wq_completion)nfc3_nci_cmd_wq#1423 ->(wq_completion)nfc2_nci_cmd_wq#1573 ->(wq_completion)nfc3_nci_cmd_wq#1424 ->(wq_completion)nfc4_nci_cmd_wq#370 ->(wq_completion)nfc2_nci_cmd_wq#1574 ->(wq_completion)nfc2_nci_cmd_wq#1575 ->(wq_completion)nfc3_nci_cmd_wq#1425 ->(wq_completion)nfc2_nci_cmd_wq#1576 ->(wq_completion)nfc3_nci_cmd_wq#1426 ->(wq_completion)nfc2_nci_cmd_wq#1577 ->(wq_completion)nfc3_nci_cmd_wq#1427 ->(wq_completion)nfc2_nci_cmd_wq#1578 ->(wq_completion)nfc3_nci_cmd_wq#1428 ->(wq_completion)nfc2_nci_cmd_wq#1579 ->(wq_completion)nfc3_nci_cmd_wq#1429 ->(wq_completion)nfc2_nci_cmd_wq#1580 ->(wq_completion)nfc3_nci_cmd_wq#1430 ->(wq_completion)nfc2_nci_cmd_wq#1581 ->(wq_completion)nfc3_nci_cmd_wq#1431 ->(wq_completion)nfc2_nci_cmd_wq#1582 ->(wq_completion)nfc3_nci_cmd_wq#1432 ->(wq_completion)nfc2_nci_cmd_wq#1583 ->(wq_completion)nfc3_nci_cmd_wq#1433 ->(wq_completion)nfc2_nci_cmd_wq#1584 ->(wq_completion)nfc3_nci_cmd_wq#1434 ->(wq_completion)nfc2_nci_cmd_wq#1585 ->(wq_completion)nfc3_nci_cmd_wq#1435 ->(wq_completion)nfc2_nci_cmd_wq#1586 ->(wq_completion)nfc3_nci_cmd_wq#1436 ->(wq_completion)nfc2_nci_cmd_wq#1587 ->(wq_completion)nfc3_nci_cmd_wq#1437 ->(wq_completion)nfc4_nci_cmd_wq#371 ->(wq_completion)nfc2_nci_cmd_wq#1588 ->(wq_completion)nfc2_nci_cmd_wq#1589 ->(wq_completion)nfc3_nci_cmd_wq#1438 ->(wq_completion)nfc2_nci_cmd_wq#1590 ->(wq_completion)nfc3_nci_cmd_wq#1439 ->(wq_completion)nfc2_nci_cmd_wq#1591 ->(wq_completion)nfc3_nci_cmd_wq#1440 ->(wq_completion)nfc2_nci_cmd_wq#1592 ->(wq_completion)nfc3_nci_cmd_wq#1441 ->(wq_completion)nfc2_nci_cmd_wq#1593 ->(wq_completion)nfc3_nci_cmd_wq#1442 ->(wq_completion)nfc2_nci_cmd_wq#1594 ->(wq_completion)nfc3_nci_cmd_wq#1443 ->(wq_completion)nfc2_nci_cmd_wq#1595 ->(wq_completion)nfc3_nci_cmd_wq#1444 ->(wq_completion)nfc2_nci_cmd_wq#1596 ->(wq_completion)nfc3_nci_cmd_wq#1445 ->(wq_completion)nfc2_nci_cmd_wq#1597 ->(wq_completion)nfc3_nci_cmd_wq#1446 ->(wq_completion)nfc2_nci_cmd_wq#1598 ->(wq_completion)nfc3_nci_cmd_wq#1447 ->(wq_completion)nfc2_nci_cmd_wq#1599 ->(wq_completion)nfc3_nci_cmd_wq#1448 ->(wq_completion)nfc2_nci_cmd_wq#1600 ->(wq_completion)nfc3_nci_cmd_wq#1449 ->(wq_completion)nfc2_nci_cmd_wq#1601 ->(wq_completion)nfc3_nci_cmd_wq#1450 ->(wq_completion)nfc2_nci_cmd_wq#1602 ->(wq_completion)nfc3_nci_cmd_wq#1451 ->(wq_completion)nfc2_nci_cmd_wq#1603 ->(wq_completion)nfc3_nci_cmd_wq#1452 ->(wq_completion)nfc2_nci_cmd_wq#1604 ->(wq_completion)nfc3_nci_cmd_wq#1453 ->(wq_completion)nfc2_nci_cmd_wq#1605 ->(wq_completion)nfc3_nci_cmd_wq#1454 ->(wq_completion)nfc2_nci_cmd_wq#1606 ->(wq_completion)nfc3_nci_cmd_wq#1455 ->(wq_completion)nfc4_nci_cmd_wq#372 ->(wq_completion)nfc2_nci_cmd_wq#1607 ->(wq_completion)nfc2_nci_cmd_wq#1608 ->(wq_completion)nfc3_nci_cmd_wq#1456 ->(wq_completion)nfc2_nci_cmd_wq#1609 ->(wq_completion)nfc3_nci_cmd_wq#1457 ->(wq_completion)nfc2_nci_cmd_wq#1610 ->(wq_completion)nfc3_nci_cmd_wq#1458 ->(wq_completion)nfc2_nci_cmd_wq#1611 ->(wq_completion)nfc3_nci_cmd_wq#1459 ->(wq_completion)nfc4_nci_cmd_wq#373 ->(wq_completion)nfc5_nci_cmd_wq#262 ->(wq_completion)nfc2_nci_cmd_wq#1612 ->(wq_completion)nfc3_nci_cmd_wq#1460 ->(wq_completion)nfc2_nci_cmd_wq#1613 ->(wq_completion)nfc3_nci_cmd_wq#1461 ->(wq_completion)nfc2_nci_cmd_wq#1614 ->(wq_completion)nfc3_nci_cmd_wq#1462 ->(wq_completion)nfc2_nci_cmd_wq#1615 ->(wq_completion)nfc3_nci_cmd_wq#1463 ->(wq_completion)nfc2_nci_cmd_wq#1616 ->(wq_completion)nfc3_nci_cmd_wq#1464 ->(wq_completion)nfc2_nci_cmd_wq#1617 ->(wq_completion)nfc3_nci_cmd_wq#1465 ->(wq_completion)nfc2_nci_cmd_wq#1618 ->(wq_completion)nfc3_nci_cmd_wq#1466 ->(wq_completion)nfc2_nci_cmd_wq#1619 ->(wq_completion)nfc3_nci_cmd_wq#1467 ->(wq_completion)nfc2_nci_cmd_wq#1620 ->(wq_completion)nfc3_nci_cmd_wq#1468 ->(wq_completion)nfc2_nci_cmd_wq#1621 ->(wq_completion)nfc3_nci_cmd_wq#1469 ->(wq_completion)nfc2_nci_cmd_wq#1622 ->(wq_completion)nfc3_nci_cmd_wq#1470 ->(wq_completion)nfc4_nci_cmd_wq#374 ->(wq_completion)nfc2_nci_cmd_wq#1623 ->(wq_completion)nfc2_nci_cmd_wq#1624 ->(wq_completion)nfc3_nci_cmd_wq#1471 ->(wq_completion)nfc2_nci_cmd_wq#1625 ->(wq_completion)nfc3_nci_cmd_wq#1472 ->(wq_completion)nfc2_nci_cmd_wq#1626 ->(wq_completion)nfc3_nci_cmd_wq#1473 ->(wq_completion)nfc2_nci_cmd_wq#1627 ->(wq_completion)nfc3_nci_cmd_wq#1474 ->(wq_completion)nfc2_nci_cmd_wq#1628 ->(wq_completion)nfc3_nci_cmd_wq#1475 ->(wq_completion)nfc2_nci_cmd_wq#1629 ->(wq_completion)nfc2_nci_cmd_wq#1631 ->(wq_completion)nfc2_nci_cmd_wq#1632 ->(wq_completion)nfc2_nci_cmd_wq#1633 ->(wq_completion)nfc2_nci_cmd_wq#1634 ->(wq_completion)nfc2_nci_cmd_wq#1635 ->(wq_completion)nfc2_nci_cmd_wq#1636 ->(wq_completion)nfc2_nci_cmd_wq#1637 ->(wq_completion)nfc2_nci_cmd_wq#1638 ->(wq_completion)nfc2_nci_cmd_wq#1639 ->(wq_completion)nfc2_nci_cmd_wq#1640 ->(wq_completion)nfc2_nci_cmd_wq#1641 ->(wq_completion)nfc2_nci_cmd_wq#1642 ->(wq_completion)nfc2_nci_cmd_wq#1643 ->(wq_completion)nfc2_nci_cmd_wq#1644 ->(wq_completion)nfc2_nci_cmd_wq#1645 ->(wq_completion)nfc3_nci_cmd_wq#1476 ->(wq_completion)nfc2_nci_cmd_wq#1646 ->(wq_completion)nfc2_nci_cmd_wq#1647 ->(wq_completion)nfc3_nci_cmd_wq#1477 ->(wq_completion)nfc2_nci_cmd_wq#1648 ->(wq_completion)nfc2_nci_cmd_wq#1649 ->(wq_completion)nfc2_nci_cmd_wq#1650 ->(wq_completion)nfc2_nci_cmd_wq#1651 ->(wq_completion)nfc2_nci_cmd_wq#1652 ->(wq_completion)nfc2_nci_cmd_wq#1653 ->(wq_completion)nfc3_nci_cmd_wq#1478 ->(wq_completion)nfc2_nci_cmd_wq#1654 ->(wq_completion)nfc2_nci_cmd_wq#1655 ->(wq_completion)nfc2_nci_cmd_wq#1656 ->(wq_completion)nfc2_nci_cmd_wq#1657 ->(wq_completion)nfc3_nci_cmd_wq#1479 ->(wq_completion)nfc2_nci_cmd_wq#1658 ->(wq_completion)nfc3_nci_cmd_wq#1480 ->(wq_completion)nfc2_nci_cmd_wq#1659 ->(wq_completion)nfc3_nci_cmd_wq#1481 ->(wq_completion)nfc2_nci_cmd_wq#1660 ->(wq_completion)nfc2_nci_cmd_wq#1661 ->(wq_completion)nfc2_nci_cmd_wq#1662 ->(wq_completion)nfc2_nci_cmd_wq#1663 ->(wq_completion)nfc2_nci_cmd_wq#1664 ->(wq_completion)nfc2_nci_cmd_wq#1665 ->(wq_completion)nfc2_nci_cmd_wq#1666 ->(wq_completion)nfc2_nci_cmd_wq#1667 ->(wq_completion)nfc2_nci_cmd_wq#1668 ->(wq_completion)nfc2_nci_cmd_wq#1669 ->(wq_completion)nfc2_nci_cmd_wq#1670 ->(wq_completion)nfc2_nci_cmd_wq#1671 ->(wq_completion)nfc2_nci_cmd_wq#1672 ->(wq_completion)nfc2_nci_cmd_wq#1673 ->(wq_completion)nfc2_nci_cmd_wq#1674 ->(wq_completion)nfc2_nci_cmd_wq#1675 ->(wq_completion)nfc2_nci_cmd_wq#1676 ->(wq_completion)nfc3_nci_cmd_wq#1482 ->(wq_completion)nfc2_nci_cmd_wq#1677 ->(wq_completion)nfc2_nci_cmd_wq#1678 ->(wq_completion)nfc2_nci_cmd_wq#1679 ->(wq_completion)nfc3_nci_cmd_wq#1483 ->(wq_completion)nfc4_nci_cmd_wq#375 ->(wq_completion)nfc5_nci_cmd_wq#263 ->(wq_completion)nfc6_nci_cmd_wq#47 ->(wq_completion)nfc7_nci_cmd_wq#36 ->(wq_completion)nfc8_nci_cmd_wq#7 ->(wq_completion)nfc9_nci_cmd_wq#2 ->(wq_completion)nfc10_nci_cmd_wq ->(wq_completion)nfc11_nci_cmd_wq ->(wq_completion)nfc12_nci_cmd_wq ->(wq_completion)nfc13_nci_cmd_wq ->(wq_completion)nfc14_nci_cmd_wq ->(wq_completion)nfc15_nci_cmd_wq ->(wq_completion)nfc16_nci_cmd_wq ->(wq_completion)nfc17_nci_cmd_wq ->(wq_completion)nfc18_nci_cmd_wq ->(wq_completion)nfc19_nci_cmd_wq ->(wq_completion)nfc20_nci_cmd_wq ->(wq_completion)nfc21_nci_cmd_wq ->(wq_completion)nfc22_nci_cmd_wq ->(wq_completion)nfc23_nci_cmd_wq ->(wq_completion)nfc24_nci_cmd_wq ->(wq_completion)nfc25_nci_cmd_wq ->(wq_completion)nfc26_nci_cmd_wq ->(wq_completion)nfc27_nci_cmd_wq ->(wq_completion)nfc2_nci_cmd_wq#1680 ->(wq_completion)nfc3_nci_cmd_wq#1484 ->(wq_completion)nfc4_nci_cmd_wq#376 ->(wq_completion)nfc5_nci_cmd_wq#264 ->(wq_completion)nfc6_nci_cmd_wq#48 ->(wq_completion)nfc7_nci_cmd_wq#37 ->(wq_completion)nfc8_nci_cmd_wq#8 ->(wq_completion)nfc9_nci_cmd_wq#3 ->(wq_completion)nfc10_nci_cmd_wq#2 ->(wq_completion)nfc11_nci_cmd_wq#2 ->(wq_completion)nfc56_nci_cmd_wq ->(wq_completion)nfc55_nci_cmd_wq ->(wq_completion)nfc54_nci_cmd_wq ->(wq_completion)nfc53_nci_cmd_wq ->(wq_completion)nfc52_nci_cmd_wq ->(wq_completion)nfc51_nci_cmd_wq ->(wq_completion)nfc50_nci_cmd_wq ->(wq_completion)nfc49_nci_cmd_wq ->(wq_completion)nfc48_nci_cmd_wq ->(wq_completion)nfc47_nci_cmd_wq ->(wq_completion)nfc46_nci_cmd_wq ->(wq_completion)nfc45_nci_cmd_wq ->(wq_completion)nfc44_nci_cmd_wq ->(wq_completion)nfc43_nci_cmd_wq ->(wq_completion)nfc42_nci_cmd_wq ->(wq_completion)nfc41_nci_cmd_wq ->(wq_completion)nfc40_nci_cmd_wq ->(wq_completion)nfc39_nci_cmd_wq ->(wq_completion)nfc38_nci_cmd_wq ->(wq_completion)nfc37_nci_cmd_wq ->(wq_completion)nfc36_nci_cmd_wq ->(wq_completion)nfc35_nci_cmd_wq ->(wq_completion)nfc34_nci_cmd_wq ->(wq_completion)nfc33_nci_cmd_wq ->(wq_completion)nfc32_nci_cmd_wq ->(wq_completion)nfc31_nci_cmd_wq ->(wq_completion)nfc30_nci_cmd_wq ->(wq_completion)nfc29_nci_cmd_wq ->(wq_completion)nfc28_nci_cmd_wq ->(wq_completion)nfc2_nci_cmd_wq#1681 ->(wq_completion)nfc3_nci_cmd_wq#1485 ->(wq_completion)nfc2_nci_cmd_wq#1683 ->(wq_completion)nfc2_nci_cmd_wq#1685 ->(wq_completion)nfc2_nci_cmd_wq#1686 ->(wq_completion)nfc2_nci_cmd_wq#1687 ->(wq_completion)nfc2_nci_cmd_wq#1688 ->(wq_completion)nfc2_nci_cmd_wq#1689 ->(wq_completion)nfc2_nci_cmd_wq#1690 ->(wq_completion)nfc2_nci_cmd_wq#1691 ->(wq_completion)nfc2_nci_cmd_wq#1692 ->(wq_completion)nfc2_nci_cmd_wq#1693 ->(wq_completion)nfc2_nci_cmd_wq#1694 ->(wq_completion)nfc2_nci_cmd_wq#1695 ->(wq_completion)nfc2_nci_cmd_wq#1696 ->(wq_completion)nfc2_nci_cmd_wq#1697 ->(wq_completion)nfc2_nci_cmd_wq#1698 ->(wq_completion)nfc3_nci_cmd_wq#1486 ->(wq_completion)nfc2_nci_cmd_wq#1700 ->(wq_completion)nfc2_nci_cmd_wq#1701 ->(wq_completion)nfc2_nci_cmd_wq#1702 ->(wq_completion)nfc3_nci_cmd_wq#1487 ->(wq_completion)nfc4_nci_cmd_wq#377 ->(wq_completion)nfc2_nci_cmd_wq#1703 ->(wq_completion)nfc3_nci_cmd_wq#1489 ->(wq_completion)nfc4_nci_cmd_wq#378 ->(wq_completion)nfc2_nci_cmd_wq#1704 ->(wq_completion)nfc3_nci_cmd_wq#1490 ->(wq_completion)nfc2_nci_cmd_wq#1705 ->(wq_completion)nfc4_nci_cmd_wq#379 ->(wq_completion)nfc3_nci_cmd_wq#1491 ->(wq_completion)nfc2_nci_cmd_wq#1706 ->(wq_completion)nfc3_nci_cmd_wq#1492 ->(wq_completion)nfc2_nci_cmd_wq#1707 ->(wq_completion)nfc2_nci_cmd_wq#1708 ->(wq_completion)nfc3_nci_cmd_wq#1493 ->(wq_completion)nfc4_nci_cmd_wq#380 ->(wq_completion)nfc2_nci_cmd_wq#1709 ->(wq_completion)nfc2_nci_cmd_wq#1710 ->(wq_completion)nfc3_nci_cmd_wq#1494 ->(wq_completion)nfc2_nci_cmd_wq#1711 ->(wq_completion)nfc2_nci_cmd_wq#1712 ->(wq_completion)nfc2_nci_cmd_wq#1713 ->(wq_completion)nfc3_nci_cmd_wq#1495 ->(wq_completion)nfc2_nci_cmd_wq#1714 ->(wq_completion)nfc3_nci_cmd_wq#1496 ->(wq_completion)nfc4_nci_cmd_wq#381 ->(wq_completion)nfc2_nci_cmd_wq#1715 ->(wq_completion)nfc2_nci_cmd_wq#1716 ->(wq_completion)nfc3_nci_cmd_wq#1497 ->(wq_completion)nfc2_nci_cmd_wq#1718 ->(wq_completion)nfc7_nci_cmd_wq#38 ->(wq_completion)nfc8_nci_cmd_wq#9 ->(wq_completion)nfc4_nci_cmd_wq#382 ->(wq_completion)nfc5_nci_cmd_wq#265 ->(wq_completion)nfc9_nci_cmd_wq#4 ->(wq_completion)nfc6_nci_cmd_wq#49 ->(wq_completion)nfc4_nci_cmd_wq#383 ->(wq_completion)nfc2_nci_cmd_wq#1719 ->(wq_completion)nfc5_nci_cmd_wq#266 ->(wq_completion)nfc3_nci_cmd_wq#1498 ->(wq_completion)nfc7_nci_cmd_wq#39 ->(wq_completion)nfc6_nci_cmd_wq#50 ->(wq_completion)nfc3_nci_cmd_wq#1499 ->(wq_completion)nfc2_nci_cmd_wq#1720 ->(wq_completion)nfc6_nci_cmd_wq#51 ->(wq_completion)nfc4_nci_cmd_wq#384 ->(wq_completion)nfc7_nci_cmd_wq#40 ->(wq_completion)nfc5_nci_cmd_wq#267 ->(wq_completion)nfc8_nci_cmd_wq#10 ->(wq_completion)nfc2_nci_cmd_wq#1721 ->(wq_completion)nfc4_nci_cmd_wq#385 ->(wq_completion)nfc3_nci_cmd_wq#1500 ->(wq_completion)nfc2_nci_cmd_wq#1722 ->(wq_completion)nfc8_nci_cmd_wq#11 ->(wq_completion)nfc5_nci_cmd_wq#268 ->(wq_completion)nfc6_nci_cmd_wq#52 ->(wq_completion)nfc9_nci_cmd_wq#5 ->(wq_completion)nfc7_nci_cmd_wq#41 ->(wq_completion)nfc4_nci_cmd_wq#386 ->(wq_completion)nfc3_nci_cmd_wq#1501 ->(wq_completion)nfc8_nci_cmd_wq#12 ->(wq_completion)nfc10_nci_cmd_wq#3 ->(wq_completion)nfc2_nci_cmd_wq#1723 ->(wq_completion)nfc11_nci_cmd_wq#3 ->(wq_completion)nfc6_nci_cmd_wq#53 ->(wq_completion)nfc5_nci_cmd_wq#269 ->(wq_completion)nfc8_nci_cmd_wq#13 ->(wq_completion)nfc2_nci_cmd_wq#1724 ->(wq_completion)nfc9_nci_cmd_wq#6 ->(wq_completion)nfc3_nci_cmd_wq#1502 ->(wq_completion)nfc4_nci_cmd_wq#387 ->(wq_completion)nfc7_nci_cmd_wq#42 ->(wq_completion)nfc11_nci_cmd_wq#4 ->(wq_completion)nfc6_nci_cmd_wq#54 ->(wq_completion)nfc8_nci_cmd_wq#14 ->(wq_completion)nfc5_nci_cmd_wq#270 ->(wq_completion)nfc2_nci_cmd_wq#1725 ->(wq_completion)nfc10_nci_cmd_wq#4 ->(wq_completion)nfc9_nci_cmd_wq#7 ->(wq_completion)nfc4_nci_cmd_wq#388 ->(wq_completion)nfc11_nci_cmd_wq#5 ->(wq_completion)nfc3_nci_cmd_wq#1503 ->(wq_completion)nfc6_nci_cmd_wq#55 ->(wq_completion)nfc7_nci_cmd_wq#43 ->(wq_completion)nfc6_nci_cmd_wq#56 ->(wq_completion)nfc3_nci_cmd_wq#1504 ->(wq_completion)nfc8_nci_cmd_wq#15 ->(wq_completion)nfc4_nci_cmd_wq#389 ->(wq_completion)nfc5_nci_cmd_wq#271 ->(wq_completion)nfc2_nci_cmd_wq#1726 ->(wq_completion)nfc10_nci_cmd_wq#5 ->(wq_completion)nfc12_nci_cmd_wq#2 ->(wq_completion)nfc7_nci_cmd_wq#44 ->(wq_completion)nfc13_nci_cmd_wq#2 ->(wq_completion)nfc9_nci_cmd_wq#8 ->(wq_completion)nfc11_nci_cmd_wq#6 ->(wq_completion)nfc4_nci_cmd_wq#390 ->(wq_completion)nfc2_nci_cmd_wq#1727 ->(wq_completion)nfc7_nci_cmd_wq#45 ->(wq_completion)nfc5_nci_cmd_wq#272 ->(wq_completion)nfc6_nci_cmd_wq#57 ->(wq_completion)nfc3_nci_cmd_wq#1505 ->(wq_completion)nfc6_nci_cmd_wq#58 ->(wq_completion)nfc10_nci_cmd_wq#6 ->(wq_completion)nfc7_nci_cmd_wq#46 ->(wq_completion)nfc2_nci_cmd_wq#1728 ->(wq_completion)nfc8_nci_cmd_wq#16 ->(wq_completion)nfc9_nci_cmd_wq#9 ->(wq_completion)nfc3_nci_cmd_wq#1506 ->(wq_completion)nfc4_nci_cmd_wq#391 ->(wq_completion)nfc11_nci_cmd_wq#7 ->(wq_completion)nfc5_nci_cmd_wq#273 ->(wq_completion)nfc4_nci_cmd_wq#392 ->(wq_completion)nfc5_nci_cmd_wq#274 ->(wq_completion)nfc2_nci_cmd_wq#1729 ->(wq_completion)nfc6_nci_cmd_wq#59 ->(wq_completion)nfc8_nci_cmd_wq#17 ->(wq_completion)nfc4_nci_cmd_wq#393 ->(wq_completion)nfc7_nci_cmd_wq#47 ->(wq_completion)nfc3_nci_cmd_wq#1507 ->(wq_completion)nfc6_nci_cmd_wq#60 ->(wq_completion)nfc7_nci_cmd_wq#48 ->(wq_completion)nfc5_nci_cmd_wq#275 ->(wq_completion)nfc3_nci_cmd_wq#1508 ->(wq_completion)nfc6_nci_cmd_wq#61 ->(wq_completion)nfc3_nci_cmd_wq#1509 ->(wq_completion)nfc4_nci_cmd_wq#394 ->(wq_completion)nfc2_nci_cmd_wq#1730 ->(wq_completion)nfc6_nci_cmd_wq#62 ->(wq_completion)nfc3_nci_cmd_wq#1510 ->(wq_completion)nfc3_nci_cmd_wq#1511 ->(wq_completion)nfc7_nci_cmd_wq#49 ->(wq_completion)nfc4_nci_cmd_wq#395 ->(wq_completion)nfc5_nci_cmd_wq#276 ->(wq_completion)nfc8_nci_cmd_wq#18 ->(wq_completion)nfc9_nci_cmd_wq#10 ->(wq_completion)nfc3_nci_cmd_wq#1512 ->(wq_completion)nfc4_nci_cmd_wq#396 ->(wq_completion)nfc6_nci_cmd_wq#63 ->(wq_completion)nfc2_nci_cmd_wq#1731 ->(wq_completion)nfc6_nci_cmd_wq#64 ->(wq_completion)nfc4_nci_cmd_wq#397 ->(wq_completion)nfc7_nci_cmd_wq#50 ->(wq_completion)nfc8_nci_cmd_wq#19 ->(wq_completion)nfc5_nci_cmd_wq#277 ->(wq_completion)nfc2_nci_cmd_wq#1732 ->(wq_completion)nfc3_nci_cmd_wq#1513 ->(wq_completion)nfc9_nci_cmd_wq#11 ->(wq_completion)nfc3_nci_cmd_wq#1514 ->(wq_completion)nfc5_nci_cmd_wq#278 ->(wq_completion)nfc3_nci_cmd_wq#1515 ->(wq_completion)nfc5_nci_cmd_wq#279 ->(wq_completion)nfc4_nci_cmd_wq#398 ->(wq_completion)nfc2_nci_cmd_wq#1733 ->(wq_completion)nfc5_nci_cmd_wq#280 ->(wq_completion)nfc6_nci_cmd_wq#65 ->(wq_completion)nfc3_nci_cmd_wq#1516 ->(wq_completion)nfc7_nci_cmd_wq#51 ->(wq_completion)nfc3_nci_cmd_wq#1517 ->(wq_completion)nfc4_nci_cmd_wq#399 ->(wq_completion)nfc4_nci_cmd_wq#400 ->(wq_completion)nfc3_nci_cmd_wq#1518 ->(wq_completion)nfc5_nci_cmd_wq#281 ->(wq_completion)nfc7_nci_cmd_wq#52 ->(wq_completion)nfc10_nci_cmd_wq#7 ->(wq_completion)nfc8_nci_cmd_wq#20 ->(wq_completion)nfc2_nci_cmd_wq#1734 ->(wq_completion)nfc9_nci_cmd_wq#12 ->(wq_completion)nfc6_nci_cmd_wq#66 ->(wq_completion)nfc4_nci_cmd_wq#401 ->(wq_completion)nfc7_nci_cmd_wq#53 ->(wq_completion)nfc5_nci_cmd_wq#282 ->(wq_completion)nfc6_nci_cmd_wq#67 ->(wq_completion)nfc3_nci_cmd_wq#1519 ->(wq_completion)nfc2_nci_cmd_wq#1735 ->(wq_completion)nfc4_nci_cmd_wq#402 ->(wq_completion)nfc6_nci_cmd_wq#68 ->(wq_completion)nfc5_nci_cmd_wq#283 ->(wq_completion)nfc4_nci_cmd_wq#403 ->(wq_completion)nfc3_nci_cmd_wq#1520 ->(wq_completion)nfc2_nci_cmd_wq#1736 ->(wq_completion)nfc4_nci_cmd_wq#404 ->(wq_completion)nfc6_nci_cmd_wq#69 ->(wq_completion)nfc2_nci_cmd_wq#1737 ->(wq_completion)nfc4_nci_cmd_wq#405 ->(wq_completion)nfc3_nci_cmd_wq#1521 ->(wq_completion)nfc2_nci_cmd_wq#1738 ->(wq_completion)nfc5_nci_cmd_wq#284 ->(wq_completion)nfc2_nci_cmd_wq#1739 ->(wq_completion)nfc4_nci_cmd_wq#406 ->(wq_completion)nfc3_nci_cmd_wq#1522 ->(wq_completion)nfc3_nci_cmd_wq#1523 ->(wq_completion)nfc4_nci_cmd_wq#407 ->(wq_completion)nfc4_nci_cmd_wq#408 ->(wq_completion)nfc5_nci_cmd_wq#285 ->(wq_completion)nfc6_nci_cmd_wq#70 ->(wq_completion)nfc4_nci_cmd_wq#409 ->(wq_completion)nfc3_nci_cmd_wq#1524 ->(wq_completion)nfc2_nci_cmd_wq#1740 ->(wq_completion)nfc2_nci_cmd_wq#1741 ->(wq_completion)nfc2_nci_cmd_wq#1742 ->(wq_completion)nfc2_nci_cmd_wq#1743 ->(wq_completion)nfc4_nci_cmd_wq#410 ->(wq_completion)nfc2_nci_cmd_wq#1744 ->(wq_completion)nfc3_nci_cmd_wq#1525 ->(wq_completion)nfc3_nci_cmd_wq#1526 ->(wq_completion)nfc2_nci_cmd_wq#1745 ->(wq_completion)nfc3_nci_cmd_wq#1527 ->(wq_completion)nfc2_nci_cmd_wq#1746 ->(wq_completion)nfc4_nci_cmd_wq#411 ->(wq_completion)nfc3_nci_cmd_wq#1528 ->(wq_completion)nfc2_nci_cmd_wq#1747 ->(wq_completion)nfc3_nci_cmd_wq#1529 ->(wq_completion)nfc2_nci_cmd_wq#1748 ->(wq_completion)nfc3_nci_cmd_wq#1530 ->(wq_completion)nfc2_nci_cmd_wq#1749 ->(wq_completion)nfc3_nci_cmd_wq#1531 ->(wq_completion)nfc2_nci_cmd_wq#1750 ->(wq_completion)nfc3_nci_cmd_wq#1532 ->(wq_completion)nfc7_nci_cmd_wq#54 ->(wq_completion)nfc9_nci_cmd_wq#13 ->(wq_completion)nfc2_nci_cmd_wq#1751 ->(wq_completion)nfc5_nci_cmd_wq#286 ->(wq_completion)nfc4_nci_cmd_wq#412 ->(wq_completion)nfc2_nci_cmd_wq#1752 ->(wq_completion)nfc3_nci_cmd_wq#1533 ->(wq_completion)nfc8_nci_cmd_wq#21 ->(wq_completion)nfc6_nci_cmd_wq#71 ->(wq_completion)nfc2_nci_cmd_wq#1753 ->(wq_completion)nfc3_nci_cmd_wq#1534 ->(wq_completion)nfc4_nci_cmd_wq#413 ->(wq_completion)nfc5_nci_cmd_wq#287 ->(wq_completion)nfc4_nci_cmd_wq#414 ->(wq_completion)nfc3_nci_cmd_wq#1535 ->(wq_completion)nfc6_nci_cmd_wq#72 ->(wq_completion)nfc2_nci_cmd_wq#1754 ->(wq_completion)nfc2_nci_cmd_wq#1755 ->(wq_completion)nfc3_nci_cmd_wq#1536 ->(wq_completion)nfc2_nci_cmd_wq#1756 ->(wq_completion)nfc4_nci_cmd_wq#415 ->(wq_completion)nfc3_nci_cmd_wq#1537 ->(wq_completion)nfc3_nci_cmd_wq#1538 ->(wq_completion)nfc2_nci_cmd_wq#1757 ->(wq_completion)nfc4_nci_cmd_wq#416 ->(wq_completion)nfc3_nci_cmd_wq#1539 ->(wq_completion)nfc2_nci_cmd_wq#1758 ->(wq_completion)nfc2_nci_cmd_wq#1759 ->(wq_completion)nfc2_nci_cmd_wq#1761 ->(wq_completion)nfc3_nci_cmd_wq#1540 ->(wq_completion)nfc4_nci_cmd_wq#417 ->(wq_completion)nfc2_nci_cmd_wq#1763 ->(wq_completion)nfc3_nci_cmd_wq#1541 ->(wq_completion)nfc4_nci_cmd_wq#418 ->(wq_completion)nfc5_nci_cmd_wq#288 ->(wq_completion)nfc2_nci_cmd_wq#1764 ->(wq_completion)nfc3_nci_cmd_wq#1542 ->(wq_completion)nfc4_nci_cmd_wq#419 ->(wq_completion)nfc2_nci_cmd_wq#1765 ->(wq_completion)nfc3_nci_cmd_wq#1543 ->(wq_completion)nfc2_nci_cmd_wq#1766 ->(wq_completion)nfc3_nci_cmd_wq#1544 ->(wq_completion)nfc5_nci_cmd_wq#289 ->(wq_completion)nfc6_nci_cmd_wq#73 ->(wq_completion)nfc4_nci_cmd_wq#420 ->(wq_completion)nfc2_nci_cmd_wq#1767 ->(wq_completion)nfc3_nci_cmd_wq#1545 ->(wq_completion)nfc4_nci_cmd_wq#421 ->(wq_completion)nfc2_nci_cmd_wq#1768 ->(wq_completion)nfc3_nci_cmd_wq#1546 ->(wq_completion)nfc4_nci_cmd_wq#422 ->(wq_completion)nfc5_nci_cmd_wq#290 ->(wq_completion)nfc6_nci_cmd_wq#74 ->(wq_completion)nfc2_nci_cmd_wq#1769 ->(wq_completion)nfc7_nci_cmd_wq#55 ->(wq_completion)nfc3_nci_cmd_wq#1547 ->(wq_completion)nfc2_nci_cmd_wq#1770 ->(wq_completion)nfc2_nci_cmd_wq#1771 ->(wq_completion)nfc2_nci_cmd_wq#1772 ->(wq_completion)nfc3_nci_cmd_wq#1548 ->(wq_completion)nfc2_nci_cmd_wq#1774 ->(wq_completion)nfc2_nci_cmd_wq#1775 ->(wq_completion)nfc2_nci_cmd_wq#1776 ->(wq_completion)nfc2_nci_cmd_wq#1777 ->(wq_completion)nfc2_nci_cmd_wq#1778 ->(wq_completion)nfc2_nci_cmd_wq#1779 ->(wq_completion)nfc2_nci_cmd_wq#1780 ->(wq_completion)nfc2_nci_cmd_wq#1781 ->(wq_completion)nfc2_nci_cmd_wq#1782 ->(wq_completion)nfc2_nci_cmd_wq#1783 ->(wq_completion)nfc2_nci_cmd_wq#1784 ->(wq_completion)nfc2_nci_cmd_wq#1785 ->(wq_completion)nfc2_nci_cmd_wq#1786 ->(wq_completion)nfc2_nci_cmd_wq#1787 ->(wq_completion)nfc2_nci_cmd_wq#1788 ->(wq_completion)nfc2_nci_cmd_wq#1789 ->(wq_completion)nfc3_nci_cmd_wq#1550 ->(wq_completion)nfc2_nci_cmd_wq#1790 ->(wq_completion)nfc2_nci_cmd_wq#1791 ->(wq_completion)nfc2_nci_cmd_wq#1792 ->(wq_completion)nfc3_nci_cmd_wq#1551 ->(wq_completion)nfc5_nci_cmd_wq#291 ->(wq_completion)nfc8_nci_cmd_wq#22 ->(wq_completion)nfc2_nci_cmd_wq#1793 ->(wq_completion)nfc3_nci_cmd_wq#1552 ->(wq_completion)nfc4_nci_cmd_wq#423 ->(wq_completion)nfc6_nci_cmd_wq#75 ->(wq_completion)nfc7_nci_cmd_wq#56 ->(wq_completion)nfc9_nci_cmd_wq#14 ->(wq_completion)nfc2_nci_cmd_wq#1794 ->(wq_completion)nfc3_nci_cmd_wq#1553 ->(wq_completion)nfc4_nci_cmd_wq#424 ->(wq_completion)nfc2_nci_cmd_wq#1795 ->(wq_completion)nfc3_nci_cmd_wq#1554 ->(wq_completion)nfc4_nci_cmd_wq#425 ->(wq_completion)nfc2_nci_cmd_wq#1796 ->(wq_completion)nfc3_nci_cmd_wq#1555 ->(wq_completion)nfc3_nci_cmd_wq#1556 ->(wq_completion)nfc5_nci_cmd_wq#292 ->(wq_completion)nfc4_nci_cmd_wq#426 ->(wq_completion)nfc2_nci_cmd_wq#1797 ->(wq_completion)nfc6_nci_cmd_wq#76 ->(wq_completion)nfc2_nci_cmd_wq#1798 ->(wq_completion)nfc3_nci_cmd_wq#1557 ->(wq_completion)nfc2_nci_cmd_wq#1799 ->(wq_completion)nfc2_nci_cmd_wq#1800 ->(wq_completion)nfc4_nci_cmd_wq#427 ->(wq_completion)nfc3_nci_cmd_wq#1558 ->(wq_completion)nfc2_nci_cmd_wq#1801 ->(wq_completion)nfc2_nci_cmd_wq#1802 ->(wq_completion)nfc2_nci_cmd_wq#1803 ->(wq_completion)nfc2_nci_cmd_wq#1804 ->(wq_completion)nfc3_nci_cmd_wq#1559 ->(wq_completion)nfc2_nci_cmd_wq#1805 ->(wq_completion)nfc2_nci_cmd_wq#1806 ->(wq_completion)nfc2_nci_cmd_wq#1807 ->(wq_completion)nfc3_nci_cmd_wq#1560 ->(wq_completion)nfc4_nci_cmd_wq#428 ->(wq_completion)nfc2_nci_cmd_wq#1808 ->(wq_completion)nfc5_nci_cmd_wq#293 ->(wq_completion)nfc2_nci_cmd_wq#1809 ->(wq_completion)nfc3_nci_cmd_wq#1561 ->(wq_completion)nfc4_nci_cmd_wq#429 ->(wq_completion)nfc2_nci_cmd_wq#1810 ->(wq_completion)nfc3_nci_cmd_wq#1562 ->(wq_completion)nfc2_nci_cmd_wq#1811 ->(wq_completion)nfc4_nci_cmd_wq#430 ->(wq_completion)nfc2_nci_cmd_wq#1812 ->(wq_completion)nfc2_nci_cmd_wq#1813 ->(wq_completion)nfc4_nci_cmd_wq#431 ->(wq_completion)nfc3_nci_cmd_wq#1563 ->(wq_completion)nfc5_nci_cmd_wq#294 ->(wq_completion)nfc2_nci_cmd_wq#1814 ->(wq_completion)nfc3_nci_cmd_wq#1564 ->(wq_completion)nfc2_nci_cmd_wq#1815 ->(wq_completion)nfc3_nci_cmd_wq#1565 ->(wq_completion)nfc2_nci_cmd_wq#1816 ->(wq_completion)nfc3_nci_cmd_wq#1566 ->(wq_completion)nfc2_nci_cmd_wq#1817 ->(wq_completion)nfc3_nci_cmd_wq#1567 ->(wq_completion)nfc2_nci_cmd_wq#1818 ->(wq_completion)nfc3_nci_cmd_wq#1568 ->(wq_completion)nfc2_nci_cmd_wq#1819 ->(wq_completion)nfc3_nci_cmd_wq#1569 ->(wq_completion)nfc2_nci_cmd_wq#1820 ->(wq_completion)nfc3_nci_cmd_wq#1570 ->(wq_completion)nfc2_nci_cmd_wq#1821 ->(wq_completion)nfc3_nci_cmd_wq#1571 ->(wq_completion)nfc2_nci_cmd_wq#1822 ->(wq_completion)nfc2_nci_cmd_wq#1823 ->(wq_completion)nfc3_nci_cmd_wq#1572 ->(wq_completion)nfc2_nci_cmd_wq#1824 ->(wq_completion)nfc3_nci_cmd_wq#1573 ->(wq_completion)nfc2_nci_cmd_wq#1825 ->(wq_completion)nfc2_nci_cmd_wq#1826 ->(wq_completion)nfc3_nci_cmd_wq#1574 ->(wq_completion)nfc2_nci_cmd_wq#1827 ->(wq_completion)nfc2_nci_cmd_wq#1828 ->(wq_completion)nfc3_nci_cmd_wq#1575 ->(wq_completion)nfc2_nci_cmd_wq#1829 ->(wq_completion)nfc4_nci_cmd_wq#432 ->(wq_completion)nfc3_nci_cmd_wq#1576 ->(wq_completion)nfc2_nci_cmd_wq#1830 ->(wq_completion)nfc4_nci_cmd_wq#433 ->(wq_completion)nfc5_nci_cmd_wq#295 ->(wq_completion)nfc2_nci_cmd_wq#1831 ->(wq_completion)nfc3_nci_cmd_wq#1577 ->(wq_completion)nfc4_nci_cmd_wq#434 ->(wq_completion)nfc2_nci_cmd_wq#1832 ->(wq_completion)nfc5_nci_cmd_wq#296 ->(wq_completion)nfc3_nci_cmd_wq#1578 ->(wq_completion)nfc2_nci_cmd_wq#1833 ->(wq_completion)nfc3_nci_cmd_wq#1579 ->(wq_completion)nfc6_nci_cmd_wq#77 ->(wq_completion)nfc4_nci_cmd_wq#435 ->(wq_completion)nfc2_nci_cmd_wq#1834 ->(wq_completion)nfc3_nci_cmd_wq#1580 ->(wq_completion)nfc4_nci_cmd_wq#436 ->(wq_completion)nfc2_nci_cmd_wq#1835 ->(wq_completion)nfc5_nci_cmd_wq#297 ->(wq_completion)nfc6_nci_cmd_wq#78 ->(wq_completion)nfc2_nci_cmd_wq#1836 ->(wq_completion)nfc3_nci_cmd_wq#1581 ->(wq_completion)nfc6_nci_cmd_wq#79 ->(wq_completion)nfc4_nci_cmd_wq#437 ->(wq_completion)nfc7_nci_cmd_wq#57 ->(wq_completion)nfc2_nci_cmd_wq#1837 ->(wq_completion)nfc5_nci_cmd_wq#298 ->(wq_completion)nfc3_nci_cmd_wq#1582 ->(wq_completion)nfc4_nci_cmd_wq#438 ->(wq_completion)nfc2_nci_cmd_wq#1838 ->(wq_completion)nfc5_nci_cmd_wq#299 ->(wq_completion)nfc3_nci_cmd_wq#1583 ->(wq_completion)nfc5_nci_cmd_wq#300 ->(wq_completion)nfc2_nci_cmd_wq#1839 ->(wq_completion)nfc4_nci_cmd_wq#439 ->(wq_completion)nfc3_nci_cmd_wq#1584 ->(wq_completion)nfc4_nci_cmd_wq#440 ->(wq_completion)nfc2_nci_cmd_wq#1840 ->(wq_completion)nfc4_nci_cmd_wq#441 ->(wq_completion)nfc5_nci_cmd_wq#301 ->(wq_completion)nfc6_nci_cmd_wq#80 ->(wq_completion)nfc2_nci_cmd_wq#1841 ->(wq_completion)nfc3_nci_cmd_wq#1585 ->(wq_completion)nfc3_nci_cmd_wq#1586 ->(wq_completion)nfc4_nci_cmd_wq#442 ->(wq_completion)nfc3_nci_cmd_wq#1587 ->(wq_completion)nfc5_nci_cmd_wq#302 ->(wq_completion)nfc6_nci_cmd_wq#81 ->(wq_completion)nfc2_nci_cmd_wq#1842 ->(wq_completion)nfc5_nci_cmd_wq#303 ->(wq_completion)nfc2_nci_cmd_wq#1843 ->(wq_completion)nfc3_nci_cmd_wq#1588 ->(wq_completion)nfc4_nci_cmd_wq#443 ->(wq_completion)nfc3_nci_cmd_wq#1589 ->(wq_completion)nfc3_nci_cmd_wq#1590 ->(wq_completion)nfc5_nci_cmd_wq#305 ->(wq_completion)nfc4_nci_cmd_wq#444 ->(wq_completion)nfc2_nci_cmd_wq#1844 ->(wq_completion)nfc6_nci_cmd_wq#82 ->(wq_completion)nfc7_nci_cmd_wq#58 ->(wq_completion)nfc6_nci_cmd_wq#83 ->(wq_completion)nfc2_nci_cmd_wq#1845 ->(wq_completion)nfc3_nci_cmd_wq#1591 ->(wq_completion)nfc4_nci_cmd_wq#445 ->(wq_completion)nfc8_nci_cmd_wq#23 ->(wq_completion)nfc5_nci_cmd_wq#306 ->(wq_completion)nfc7_nci_cmd_wq#59 ->(wq_completion)nfc2_nci_cmd_wq#1846 ->(wq_completion)nfc3_nci_cmd_wq#1592 ->(wq_completion)nfc4_nci_cmd_wq#446 ->(wq_completion)nfc7_nci_cmd_wq#60 ->(wq_completion)nfc5_nci_cmd_wq#307 ->(wq_completion)nfc6_nci_cmd_wq#84 ->(wq_completion)nfc7_nci_cmd_wq#61 ->(wq_completion)nfc5_nci_cmd_wq#308 ->(wq_completion)nfc2_nci_cmd_wq#1847 ->(wq_completion)nfc6_nci_cmd_wq#85 ->(wq_completion)nfc4_nci_cmd_wq#447 ->(wq_completion)nfc2_nci_cmd_wq#1848 ->(wq_completion)nfc4_nci_cmd_wq#448 ->(wq_completion)nfc3_nci_cmd_wq#1593 ->(wq_completion)nfc8_nci_cmd_wq#24 ->(wq_completion)nfc4_nci_cmd_wq#449 ->(wq_completion)nfc3_nci_cmd_wq#1594 ->(wq_completion)nfc5_nci_cmd_wq#309 ->(wq_completion)nfc7_nci_cmd_wq#62 ->(wq_completion)nfc3_nci_cmd_wq#1595 ->(wq_completion)nfc4_nci_cmd_wq#450 ->(wq_completion)nfc3_nci_cmd_wq#1596 ->(wq_completion)nfc6_nci_cmd_wq#86 ->(wq_completion)nfc2_nci_cmd_wq#1849 ->(wq_completion)nfc3_nci_cmd_wq#1597 ->(wq_completion)nfc2_nci_cmd_wq#1850 ->(wq_completion)nfc3_nci_cmd_wq#1598 ->(wq_completion)nfc5_nci_cmd_wq#310 ->(wq_completion)nfc4_nci_cmd_wq#451 ->(wq_completion)nfc3_nci_cmd_wq#1599 ->(wq_completion)nfc4_nci_cmd_wq#452 ->(wq_completion)nfc6_nci_cmd_wq#87 ->(wq_completion)nfc3_nci_cmd_wq#1600 ->(wq_completion)nfc5_nci_cmd_wq#311 ->(wq_completion)nfc2_nci_cmd_wq#1851 ->(wq_completion)nfc4_nci_cmd_wq#453 ->(wq_completion)nfc7_nci_cmd_wq#63 ->(wq_completion)nfc2_nci_cmd_wq#1852 ->(wq_completion)nfc2_nci_cmd_wq#1853 ->(wq_completion)nfc4_nci_cmd_wq#454 ->(wq_completion)nfc3_nci_cmd_wq#1601 ->(wq_completion)nfc7_nci_cmd_wq#64 ->(wq_completion)nfc6_nci_cmd_wq#88 ->(wq_completion)nfc5_nci_cmd_wq#312 ->(wq_completion)nfc3_nci_cmd_wq#1602 ->(wq_completion)nfc4_nci_cmd_wq#455 ->(wq_completion)nfc2_nci_cmd_wq#1854 ->(wq_completion)nfc5_nci_cmd_wq#313 ->(wq_completion)nfc7_nci_cmd_wq#65 ->(wq_completion)nfc2_nci_cmd_wq#1855 ->(wq_completion)nfc8_nci_cmd_wq#28 ->(wq_completion)nfc10_nci_cmd_wq#8 ->(wq_completion)nfc4_nci_cmd_wq#456 ->(wq_completion)nfc3_nci_cmd_wq#1603 ->(wq_completion)nfc9_nci_cmd_wq#15 ->(wq_completion)nfc6_nci_cmd_wq#89 ->(wq_completion)nfc3_nci_cmd_wq#1604 ->(wq_completion)nfc2_nci_cmd_wq#1856 ->(wq_completion)nfc5_nci_cmd_wq#314 ->(wq_completion)nfc4_nci_cmd_wq#457 ->(wq_completion)nfc2_nci_cmd_wq#1857 ->(wq_completion)nfc4_nci_cmd_wq#458 ->(wq_completion)nfc5_nci_cmd_wq#315 ->(wq_completion)nfc3_nci_cmd_wq#1605 ->(wq_completion)nfc2_nci_cmd_wq#1858 ->(wq_completion)nfc4_nci_cmd_wq#459 ->(wq_completion)nfc3_nci_cmd_wq#1606 ->(wq_completion)nfc5_nci_cmd_wq#316 ->(wq_completion)nfc6_nci_cmd_wq#90 ->(wq_completion)nfc2_nci_cmd_wq#1859 ->(wq_completion)nfc4_nci_cmd_wq#460 ->(wq_completion)nfc3_nci_cmd_wq#1607 ->(wq_completion)nfc2_nci_cmd_wq#1860 ->(wq_completion)nfc3_nci_cmd_wq#1608 ->(wq_completion)nfc4_nci_cmd_wq#461 ->(wq_completion)nfc3_nci_cmd_wq#1609 ->(wq_completion)nfc2_nci_cmd_wq#1861 ->(wq_completion)nfc5_nci_cmd_wq#317 ->(wq_completion)nfc3_nci_cmd_wq#1610 ->(wq_completion)nfc2_nci_cmd_wq#1862 ->(wq_completion)nfc3_nci_cmd_wq#1611 ->(wq_completion)nfc5_nci_cmd_wq#318 ->(wq_completion)nfc7_nci_cmd_wq#66 ->(wq_completion)nfc4_nci_cmd_wq#462 ->(wq_completion)nfc6_nci_cmd_wq#91 ->(wq_completion)nfc2_nci_cmd_wq#1863 ->(wq_completion)nfc3_nci_cmd_wq#1612 ->(wq_completion)nfc2_nci_cmd_wq#1864 ->(wq_completion)nfc3_nci_cmd_wq#1613 ->(wq_completion)nfc5_nci_cmd_wq#319 ->(wq_completion)nfc4_nci_cmd_wq#463 ->(wq_completion)nfc2_nci_cmd_wq#1865 ->(wq_completion)nfc3_nci_cmd_wq#1614 ->(wq_completion)nfc2_nci_cmd_wq#1866 ->(wq_completion)nfc3_nci_cmd_wq#1615 ->(wq_completion)nfc4_nci_cmd_wq#464 ->(wq_completion)nfc2_nci_cmd_wq#1867 ->(wq_completion)nfc2_nci_cmd_wq#1868 ->(wq_completion)nfc3_nci_cmd_wq#1616 ->(wq_completion)nfc4_nci_cmd_wq#465 ->(wq_completion)nfc8_nci_cmd_wq#29 ->(wq_completion)nfc5_nci_cmd_wq#320 ->(wq_completion)nfc6_nci_cmd_wq#92 ->(wq_completion)nfc7_nci_cmd_wq#67 ->(wq_completion)nfc2_nci_cmd_wq#1869 ->(wq_completion)nfc2_nci_cmd_wq#1870 ->(wq_completion)nfc3_nci_cmd_wq#1617 ->(wq_completion)nfc2_nci_cmd_wq#1871 ->(wq_completion)nfc2_nci_cmd_wq#1872 ->(wq_completion)nfc2_nci_cmd_wq#1873 ->(wq_completion)nfc2_nci_cmd_wq#1874 ->(wq_completion)nfc2_nci_cmd_wq#1875 ->(wq_completion)nfc2_nci_cmd_wq#1876 ->(wq_completion)nfc2_nci_cmd_wq#1877 ->(wq_completion)nfc2_nci_cmd_wq#1878 ->(wq_completion)nfc2_nci_cmd_wq#1879 ->(wq_completion)nfc2_nci_cmd_wq#1880 ->(wq_completion)nfc2_nci_cmd_wq#1881 ->(wq_completion)nfc2_nci_cmd_wq#1882 ->(wq_completion)nfc2_nci_cmd_wq#1883 ->(wq_completion)nfc2_nci_cmd_wq#1884 ->(wq_completion)nfc2_nci_cmd_wq#1885 ->(wq_completion)nfc2_nci_cmd_wq#1886 ->(wq_completion)nfc2_nci_cmd_wq#1887 ->(wq_completion)nfc2_nci_cmd_wq#1888 ->(wq_completion)nfc2_nci_cmd_wq#1889 ->(wq_completion)nfc2_nci_cmd_wq#1891 ->(wq_completion)nfc3_nci_cmd_wq#1618 ->(wq_completion)nfc4_nci_cmd_wq#466 ->(wq_completion)nfc2_nci_cmd_wq#1892 ->(wq_completion)nfc3_nci_cmd_wq#1619 ->(wq_completion)nfc4_nci_cmd_wq#467 ->(wq_completion)nfc6_nci_cmd_wq#93 ->(wq_completion)nfc5_nci_cmd_wq#321 ->(wq_completion)nfc2_nci_cmd_wq#1893 ->(wq_completion)nfc3_nci_cmd_wq#1620 ->(wq_completion)nfc2_nci_cmd_wq#1894 ->(wq_completion)nfc2_nci_cmd_wq#1895 ->(wq_completion)nfc2_nci_cmd_wq#1896 ->(wq_completion)nfc2_nci_cmd_wq#1898 ->(wq_completion)nfc3_nci_cmd_wq#1621 ->(wq_completion)nfc2_nci_cmd_wq#1899 ->(wq_completion)nfc3_nci_cmd_wq#1622 ->(wq_completion)nfc2_nci_cmd_wq#1900 ->(wq_completion)nfc3_nci_cmd_wq#1623 ->(wq_completion)nfc2_nci_cmd_wq#1901 ->(wq_completion)nfc3_nci_cmd_wq#1624 ->(wq_completion)nfc2_nci_cmd_wq#1902 ->(wq_completion)nfc3_nci_cmd_wq#1625 ->(wq_completion)nfc2_nci_cmd_wq#1903 ->(wq_completion)nfc3_nci_cmd_wq#1626 ->(wq_completion)nfc2_nci_cmd_wq#1905 ->(wq_completion)nfc2_nci_cmd_wq#1906 ->(wq_completion)nfc2_nci_cmd_wq#1907 ->(wq_completion)nfc2_nci_cmd_wq#1908 ->(wq_completion)nfc3_nci_cmd_wq#1627 ->(wq_completion)nfc4_nci_cmd_wq#468 ->(wq_completion)nfc5_nci_cmd_wq#322 ->(wq_completion)nfc2_nci_cmd_wq#1909 ->(wq_completion)nfc3_nci_cmd_wq#1628 ->(wq_completion)nfc4_nci_cmd_wq#469 ->(wq_completion)nfc2_nci_cmd_wq#1910 ->(wq_completion)nfc3_nci_cmd_wq#1629 ->(wq_completion)nfc2_nci_cmd_wq#1911 ->(wq_completion)nfc3_nci_cmd_wq#1630 ->(wq_completion)nfc2_nci_cmd_wq#1912 ->(wq_completion)nfc3_nci_cmd_wq#1631 ->(wq_completion)nfc3_nci_cmd_wq#1632 ->(wq_completion)nfc2_nci_cmd_wq#1913 ->(wq_completion)nfc2_nci_cmd_wq#1914 ->(wq_completion)nfc3_nci_cmd_wq#1633 ->(wq_completion)nfc2_nci_cmd_wq#1915 ->(wq_completion)nfc3_nci_cmd_wq#1634 ->(wq_completion)nfc2_nci_cmd_wq#1916 ->(wq_completion)nfc3_nci_cmd_wq#1635 ->(wq_completion)nfc4_nci_cmd_wq#471 ->(wq_completion)nfc2_nci_cmd_wq#1917 ->(wq_completion)nfc3_nci_cmd_wq#1636 ->(wq_completion)nfc4_nci_cmd_wq#472 ->(wq_completion)nfc2_nci_cmd_wq#1918 ->(wq_completion)nfc3_nci_cmd_wq#1637 ->(wq_completion)nfc4_nci_cmd_wq#473 ->(wq_completion)nfc2_nci_cmd_wq#1919 ->(wq_completion)nfc3_nci_cmd_wq#1638 ->(wq_completion)nfc4_nci_cmd_wq#474 ->(wq_completion)nfc5_nci_cmd_wq#323 ->(wq_completion)nfc2_nci_cmd_wq#1920 ->(wq_completion)nfc2_nci_cmd_wq#1921 ->(wq_completion)nfc3_nci_cmd_wq#1639 ->(wq_completion)nfc4_nci_cmd_wq#475 ->(wq_completion)nfc2_nci_cmd_wq#1922 ->(wq_completion)nfc3_nci_cmd_wq#1640 ->(wq_completion)nfc4_nci_cmd_wq#476 ->(wq_completion)nfc5_nci_cmd_wq#324 ->(wq_completion)nfc2_nci_cmd_wq#1923 ->(wq_completion)nfc3_nci_cmd_wq#1641 ->(wq_completion)nfc4_nci_cmd_wq#477 ->(wq_completion)nfc5_nci_cmd_wq#325 ->(wq_completion)nfc3_nci_cmd_wq#1642 ->(wq_completion)nfc2_nci_cmd_wq#1925 ->(wq_completion)nfc4_nci_cmd_wq#478 ->(wq_completion)nfc5_nci_cmd_wq#326 ->(wq_completion)nfc6_nci_cmd_wq#94 ->(wq_completion)nfc5_nci_cmd_wq#327 ->(wq_completion)nfc4_nci_cmd_wq#479 ->(wq_completion)nfc3_nci_cmd_wq#1643 ->(wq_completion)nfc2_nci_cmd_wq#1926 ->(wq_completion)nfc6_nci_cmd_wq#96 ->(wq_completion)nfc5_nci_cmd_wq#328 ->(wq_completion)nfc2_nci_cmd_wq#1927 ->(wq_completion)nfc3_nci_cmd_wq#1644 ->(wq_completion)nfc2_nci_cmd_wq#1928 ->(wq_completion)nfc3_nci_cmd_wq#1645 ->(wq_completion)nfc2_nci_cmd_wq#1929 ->(wq_completion)nfc3_nci_cmd_wq#1646 ->(wq_completion)nfc2_nci_cmd_wq#1930 ->(wq_completion)nfc3_nci_cmd_wq#1647 ->(wq_completion)nfc2_nci_cmd_wq#1931 ->(wq_completion)nfc3_nci_cmd_wq#1648 ->(wq_completion)nfc2_nci_cmd_wq#1932 ->(wq_completion)nfc3_nci_cmd_wq#1649 ->(wq_completion)nfc4_nci_cmd_wq#480 ->(wq_completion)nfc2_nci_cmd_wq#1933 ->(wq_completion)nfc5_nci_cmd_wq#329 ->(wq_completion)nfc3_nci_cmd_wq#1650 ->(wq_completion)nfc6_nci_cmd_wq#97 ->(wq_completion)nfc4_nci_cmd_wq#481 ->(wq_completion)nfc2_nci_cmd_wq#1934 ->(wq_completion)nfc3_nci_cmd_wq#1651 ->(wq_completion)nfc4_nci_cmd_wq#482 ->(wq_completion)nfc2_nci_cmd_wq#1935 ->(wq_completion)nfc2_nci_cmd_wq#1936 ->(wq_completion)nfc5_nci_cmd_wq#330 ->(wq_completion)nfc3_nci_cmd_wq#1652 ->(wq_completion)nfc4_nci_cmd_wq#483 ->(wq_completion)nfc7_nci_cmd_wq#68 ->(wq_completion)nfc6_nci_cmd_wq#98 ->(wq_completion)nfc2_nci_cmd_wq#1937 ->(wq_completion)nfc3_nci_cmd_wq#1653 ->(wq_completion)nfc4_nci_cmd_wq#484 ->(wq_completion)nfc2_nci_cmd_wq#1938 ->(wq_completion)nfc2_nci_cmd_wq#1939 ->(wq_completion)nfc3_nci_cmd_wq#1654 ->(wq_completion)nfc2_nci_cmd_wq#1940 ->(wq_completion)nfc3_nci_cmd_wq#1655 ->(wq_completion)nfc2_nci_cmd_wq#1941 ->(wq_completion)nfc3_nci_cmd_wq#1656 ->(wq_completion)nfc6_nci_cmd_wq#99 ->(wq_completion)nfc4_nci_cmd_wq#485 ->(wq_completion)nfc10_nci_cmd_wq#9 ->(wq_completion)nfc2_nci_cmd_wq#1942 ->(wq_completion)nfc6_nci_cmd_wq#100 ->(wq_completion)nfc3_nci_cmd_wq#1657 ->(wq_completion)nfc5_nci_cmd_wq#331 ->(wq_completion)nfc7_nci_cmd_wq#69 ->(wq_completion)nfc8_nci_cmd_wq#30 ->(wq_completion)nfc9_nci_cmd_wq#16 ->(wq_completion)nfc2_nci_cmd_wq#1943 ->(wq_completion)nfc3_nci_cmd_wq#1658 ->(wq_completion)nfc4_nci_cmd_wq#486 ->(wq_completion)nfc3_nci_cmd_wq#1659 ->(wq_completion)nfc2_nci_cmd_wq#1944 ->(wq_completion)nfc4_nci_cmd_wq#487 ->(wq_completion)nfc5_nci_cmd_wq#332 ->(wq_completion)nfc2_nci_cmd_wq#1947 ->(wq_completion)nfc4_nci_cmd_wq#488 ->(wq_completion)nfc3_nci_cmd_wq#1661 ->(wq_completion)nfc3_nci_cmd_wq#1662 ->(wq_completion)nfc5_nci_cmd_wq#333 ->(wq_completion)nfc4_nci_cmd_wq#489 ->(wq_completion)nfc3_nci_cmd_wq#1663 ->(wq_completion)nfc4_nci_cmd_wq#490 ->(wq_completion)nfc2_nci_cmd_wq#1948 ->(wq_completion)nfc4_nci_cmd_wq#491 ->(wq_completion)nfc4_nci_cmd_wq#492 ->(wq_completion)nfc8_nci_cmd_wq#31 ->(wq_completion)nfc9_nci_cmd_wq#17 ->(wq_completion)nfc7_nci_cmd_wq#70 ->(wq_completion)nfc3_nci_cmd_wq#1664 ->(wq_completion)nfc3_nci_cmd_wq#1665 ->(wq_completion)nfc9_nci_cmd_wq#18 ->(wq_completion)nfc9_nci_cmd_wq#19 ->(wq_completion)nfc8_nci_cmd_wq#32 ->(wq_completion)nfc11_nci_cmd_wq#8 ->(wq_completion)nfc12_nci_cmd_wq#3 ->(wq_completion)nfc13_nci_cmd_wq#3 ->(wq_completion)nfc11_nci_cmd_wq#9 ->(wq_completion)nfc13_nci_cmd_wq#4 ->(wq_completion)nfc14_nci_cmd_wq#2 ->(wq_completion)nfc15_nci_cmd_wq#2 ->(wq_completion)nfc5_nci_cmd_wq#334 ->(wq_completion)nfc12_nci_cmd_wq#4 ->(wq_completion)nfc15_nci_cmd_wq#4 ->(wq_completion)nfc14_nci_cmd_wq#3 ->(wq_completion)nfc13_nci_cmd_wq#5 ->(wq_completion)nfc16_nci_cmd_wq#2 ->(wq_completion)nfc11_nci_cmd_wq#10 ->(wq_completion)nfc8_nci_cmd_wq#33 ->(wq_completion)nfc9_nci_cmd_wq#20 ->(wq_completion)nfc5_nci_cmd_wq#335 ->(wq_completion)nfc3_nci_cmd_wq#1666 ->(wq_completion)nfc7_nci_cmd_wq#71 ->(wq_completion)nfc4_nci_cmd_wq#493 ->(wq_completion)nfc10_nci_cmd_wq#10 ->(wq_completion)nfc6_nci_cmd_wq#101 ->(wq_completion)nfc2_nci_cmd_wq#1949 ->(wq_completion)nfc4_nci_cmd_wq#494 ->(wq_completion)nfc8_nci_cmd_wq#34 ->(wq_completion)nfc9_nci_cmd_wq#21 ->(wq_completion)nfc3_nci_cmd_wq#1667 ->(wq_completion)nfc2_nci_cmd_wq#1950 ->(wq_completion)nfc5_nci_cmd_wq#336 ->(wq_completion)nfc4_nci_cmd_wq#495 ->(wq_completion)nfc6_nci_cmd_wq#102 ->(wq_completion)nfc8_nci_cmd_wq#35 ->(wq_completion)nfc11_nci_cmd_wq#11 ->(wq_completion)nfc3_nci_cmd_wq#1668 ->(wq_completion)nfc10_nci_cmd_wq#11 ->(wq_completion)nfc5_nci_cmd_wq#337 ->(wq_completion)nfc6_nci_cmd_wq#103 ->(wq_completion)nfc7_nci_cmd_wq#72 ->(wq_completion)nfc2_nci_cmd_wq#1951 ->(wq_completion)nfc2_nci_cmd_wq#1952 ->(wq_completion)nfc9_nci_cmd_wq#22 ->(wq_completion)nfc4_nci_cmd_wq#496 ->(wq_completion)nfc4_nci_cmd_wq#497 ->(wq_completion)nfc2_nci_cmd_wq#1953 ->(wq_completion)nfc5_nci_cmd_wq#338 ->(wq_completion)nfc3_nci_cmd_wq#1669 ->(wq_completion)nfc6_nci_cmd_wq#104 ->(wq_completion)nfc14_nci_cmd_wq#4 ->(wq_completion)nfc7_nci_cmd_wq#73 ->(wq_completion)nfc2_nci_cmd_wq#1954 ->(wq_completion)nfc8_nci_cmd_wq#36 ->(wq_completion)nfc9_nci_cmd_wq#24 ->(wq_completion)nfc4_nci_cmd_wq#500 ->(wq_completion)nfc11_nci_cmd_wq#12 ->(wq_completion)nfc13_nci_cmd_wq#6 ->(wq_completion)nfc10_nci_cmd_wq#12 ->(wq_completion)nfc12_nci_cmd_wq#5 ->(wq_completion)nfc2_nci_cmd_wq#1955 ->(wq_completion)nfc3_nci_cmd_wq#1670 ->(wq_completion)nfc4_nci_cmd_wq#501 ->(wq_completion)nfc5_nci_cmd_wq#339 ->(wq_completion)nfc4_nci_cmd_wq#502 ->(wq_completion)nfc5_nci_cmd_wq#340 ->(wq_completion)nfc7_nci_cmd_wq#74 ->(wq_completion)nfc2_nci_cmd_wq#1956 ->(wq_completion)nfc2_nci_cmd_wq#1957 ->(wq_completion)nfc3_nci_cmd_wq#1671 ->(wq_completion)nfc5_nci_cmd_wq#341 ->(wq_completion)nfc6_nci_cmd_wq#105 ->(wq_completion)nfc7_nci_cmd_wq#75 ->(wq_completion)nfc8_nci_cmd_wq#37 ->(wq_completion)nfc3_nci_cmd_wq#1672 ->(wq_completion)nfc5_nci_cmd_wq#342 ->(wq_completion)nfc4_nci_cmd_wq#503 ->(wq_completion)nfc3_nci_cmd_wq#1673 ->(wq_completion)nfc4_nci_cmd_wq#504 ->(wq_completion)nfc3_nci_cmd_wq#1674 ->(wq_completion)nfc5_nci_cmd_wq#343 ->(wq_completion)nfc6_nci_cmd_wq#106 ->(wq_completion)nfc2_nci_cmd_wq#1958 ->(wq_completion)nfc7_nci_cmd_wq#76 ->(wq_completion)nfc8_nci_cmd_wq#38 ->(wq_completion)nfc2_nci_cmd_wq#1959 ->(wq_completion)nfc2_nci_cmd_wq#1960 ->(wq_completion)nfc3_nci_cmd_wq#1675 ->(wq_completion)nfc2_nci_cmd_wq#1961 ->(wq_completion)nfc2_nci_cmd_wq#1962 ->(wq_completion)nfc4_nci_cmd_wq#505 ->(wq_completion)nfc3_nci_cmd_wq#1676 ->(wq_completion)nfc2_nci_cmd_wq#1963 ->(wq_completion)nfc4_nci_cmd_wq#506 ->(wq_completion)nfc2_nci_cmd_wq#1964 ->(wq_completion)nfc4_nci_cmd_wq#507 ->(wq_completion)nfc3_nci_cmd_wq#1677 ->(wq_completion)nfc2_nci_cmd_wq#1965 ->(wq_completion)nfc3_nci_cmd_wq#1678 ->(wq_completion)nfc2_nci_cmd_wq#1966 ->(wq_completion)nfc3_nci_cmd_wq#1679 ->(wq_completion)nfc2_nci_cmd_wq#1967 ->(wq_completion)nfc3_nci_cmd_wq#1680 ->(wq_completion)nfc2_nci_cmd_wq#1968 ->(wq_completion)nfc3_nci_cmd_wq#1681 ->(wq_completion)nfc4_nci_cmd_wq#508 ->(wq_completion)nfc2_nci_cmd_wq#1969 ->(wq_completion)nfc2_nci_cmd_wq#1970 ->(wq_completion)nfc3_nci_cmd_wq#1682 ->(wq_completion)nfc5_nci_cmd_wq#344 ->(wq_completion)nfc4_nci_cmd_wq#509 ->(wq_completion)nfc6_nci_cmd_wq#107 ->(wq_completion)nfc2_nci_cmd_wq#1971 ->(wq_completion)nfc3_nci_cmd_wq#1683 ->(wq_completion)nfc4_nci_cmd_wq#510 ->(wq_completion)nfc5_nci_cmd_wq#345 ->(wq_completion)nfc6_nci_cmd_wq#108 ->(wq_completion)nfc2_nci_cmd_wq#1972 ->(wq_completion)nfc3_nci_cmd_wq#1684 ->(wq_completion)nfc5_nci_cmd_wq#346 ->(wq_completion)nfc6_nci_cmd_wq#109 ->(wq_completion)nfc3_nci_cmd_wq#1685 ->(wq_completion)nfc5_nci_cmd_wq#347 ->(wq_completion)nfc3_nci_cmd_wq#1686 ->(wq_completion)nfc5_nci_cmd_wq#348 ->(wq_completion)nfc3_nci_cmd_wq#1687 ->(wq_completion)nfc6_nci_cmd_wq#110 ->(wq_completion)nfc5_nci_cmd_wq#349 ->(wq_completion)nfc7_nci_cmd_wq#77 ->(wq_completion)nfc3_nci_cmd_wq#1688 ->(wq_completion)nfc5_nci_cmd_wq#350 ->(wq_completion)nfc4_nci_cmd_wq#511 ->(wq_completion)nfc2_nci_cmd_wq#1973 ->(wq_completion)nfc2_nci_cmd_wq#1974 ->(wq_completion)nfc3_nci_cmd_wq#1689 ->(wq_completion)nfc3_nci_cmd_wq#1690 ->(wq_completion)nfc6_nci_cmd_wq#111 ->(wq_completion)nfc3_nci_cmd_wq#1691 ->(wq_completion)nfc6_nci_cmd_wq#112 ->(wq_completion)nfc4_nci_cmd_wq#512 ->(wq_completion)nfc5_nci_cmd_wq#351 ->(wq_completion)nfc2_nci_cmd_wq#1975 ->(wq_completion)nfc2_nci_cmd_wq#1976 ->(wq_completion)nfc3_nci_cmd_wq#1692 ->(wq_completion)nfc4_nci_cmd_wq#513 ->(wq_completion)nfc2_nci_cmd_wq#1977 ->(wq_completion)nfc2_nci_cmd_wq#1978 ->(wq_completion)nfc3_nci_cmd_wq#1693 ->(wq_completion)nfc2_nci_cmd_wq#1979 ->(wq_completion)nfc3_nci_cmd_wq#1694 ->(wq_completion)nfc2_nci_cmd_wq#1980 ->(wq_completion)nfc3_nci_cmd_wq#1695 ->(wq_completion)nfc4_nci_cmd_wq#514 ->(wq_completion)nfc2_nci_cmd_wq#1981 ->(wq_completion)nfc2_nci_cmd_wq#1982 ->(wq_completion)nfc3_nci_cmd_wq#1696 ->(wq_completion)nfc4_nci_cmd_wq#515 ->(wq_completion)nfc5_nci_cmd_wq#352 ->(wq_completion)nfc6_nci_cmd_wq#113 ->(wq_completion)nfc7_nci_cmd_wq#78 ->(wq_completion)nfc2_nci_cmd_wq#1983 ->(wq_completion)nfc4_nci_cmd_wq#516 ->(wq_completion)nfc5_nci_cmd_wq#353 ->(wq_completion)nfc6_nci_cmd_wq#114 ->(wq_completion)nfc3_nci_cmd_wq#1697 ->(wq_completion)nfc4_nci_cmd_wq#518 ->(wq_completion)nfc5_nci_cmd_wq#355 ->(wq_completion)nfc2_nci_cmd_wq#1984 ->(wq_completion)nfc3_nci_cmd_wq#1698 ->(wq_completion)nfc4_nci_cmd_wq#519 ->(wq_completion)nfc5_nci_cmd_wq#356 ->(wq_completion)nfc2_nci_cmd_wq#1985 ->(wq_completion)nfc3_nci_cmd_wq#1699 ->(wq_completion)nfc2_nci_cmd_wq#1986 ->(wq_completion)nfc4_nci_cmd_wq#520 ->(wq_completion)nfc3_nci_cmd_wq#1700 ->(wq_completion)nfc5_nci_cmd_wq#357 ->(wq_completion)nfc6_nci_cmd_wq#115 ->(wq_completion)nfc2_nci_cmd_wq#1987 ->(wq_completion)nfc3_nci_cmd_wq#1701 ->(wq_completion)nfc2_nci_cmd_wq#1988 ->(wq_completion)nfc4_nci_cmd_wq#521 ->(wq_completion)nfc3_nci_cmd_wq#1702 ->(wq_completion)nfc5_nci_cmd_wq#358 ->(wq_completion)nfc2_nci_cmd_wq#1989 ->(wq_completion)nfc3_nci_cmd_wq#1703 ->(wq_completion)nfc2_nci_cmd_wq#1990 ->(wq_completion)nfc3_nci_cmd_wq#1704 ->(wq_completion)nfc4_nci_cmd_wq#522 ->(wq_completion)nfc2_nci_cmd_wq#1991 ->(wq_completion)nfc3_nci_cmd_wq#1705 ->(wq_completion)nfc2_nci_cmd_wq#1992 ->(wq_completion)nfc4_nci_cmd_wq#523 ->(wq_completion)nfc5_nci_cmd_wq#359 ->(wq_completion)nfc3_nci_cmd_wq#1706 ->(wq_completion)nfc2_nci_cmd_wq#1993 ->(wq_completion)nfc3_nci_cmd_wq#1707 ->(wq_completion)nfc2_nci_cmd_wq#1994 ->(wq_completion)nfc3_nci_cmd_wq#1708 ->(wq_completion)nfc2_nci_cmd_wq#1995 ->(wq_completion)nfc4_nci_cmd_wq#524 ->(wq_completion)nfc3_nci_cmd_wq#1709 ->(wq_completion)nfc5_nci_cmd_wq#360 ->(wq_completion)nfc2_nci_cmd_wq#1996 ->(wq_completion)nfc3_nci_cmd_wq#1710 ->(wq_completion)nfc3_nci_cmd_wq#1711 ->(wq_completion)nfc2_nci_cmd_wq#1997 ->(wq_completion)nfc2_nci_cmd_wq#1998 ->(wq_completion)nfc3_nci_cmd_wq#1712 ->(wq_completion)nfc2_nci_cmd_wq#1999 ->(wq_completion)nfc3_nci_cmd_wq#1713 ->(wq_completion)nfc4_nci_cmd_wq#527 ->(wq_completion)nfc5_nci_cmd_wq#361 ->(wq_completion)nfc6_nci_cmd_wq#116 ->(wq_completion)nfc6_nci_cmd_wq#117 ->(wq_completion)nfc2_nci_cmd_wq#2000 ->(wq_completion)nfc3_nci_cmd_wq#1716 ->(wq_completion)nfc4_nci_cmd_wq#528 ->(wq_completion)nfc5_nci_cmd_wq#362 ->(wq_completion)nfc7_nci_cmd_wq#79 ->(wq_completion)nfc2_nci_cmd_wq#2001 ->(wq_completion)nfc4_nci_cmd_wq#529 ->(wq_completion)nfc6_nci_cmd_wq#118 ->(wq_completion)nfc2_nci_cmd_wq#2002 ->(wq_completion)nfc3_nci_cmd_wq#1717 ->(wq_completion)nfc5_nci_cmd_wq#363 ->(wq_completion)nfc4_nci_cmd_wq#530 ->(wq_completion)nfc7_nci_cmd_wq#80 ->(wq_completion)nfc2_nci_cmd_wq#2003 ->(wq_completion)nfc3_nci_cmd_wq#1718 ->(wq_completion)nfc4_nci_cmd_wq#531 ->(wq_completion)nfc6_nci_cmd_wq#119 ->(wq_completion)nfc7_nci_cmd_wq#81 ->(wq_completion)nfc2_nci_cmd_wq#2004 ->(wq_completion)nfc3_nci_cmd_wq#1719 ->(wq_completion)nfc4_nci_cmd_wq#532 ->(wq_completion)nfc5_nci_cmd_wq#364 ->(wq_completion)nfc8_nci_cmd_wq#39 ->(wq_completion)nfc4_nci_cmd_wq#534 ->(wq_completion)nfc3_nci_cmd_wq#1720 ->(wq_completion)nfc5_nci_cmd_wq#365 ->(wq_completion)nfc2_nci_cmd_wq#2005 ->(wq_completion)nfc2_nci_cmd_wq#2006 ->(wq_completion)nfc4_nci_cmd_wq#535 ->(wq_completion)nfc3_nci_cmd_wq#1721 ->(wq_completion)nfc5_nci_cmd_wq#366 ->(wq_completion)nfc2_nci_cmd_wq#2007 ->(wq_completion)nfc3_nci_cmd_wq#1722 ->(wq_completion)nfc4_nci_cmd_wq#536 ->(wq_completion)nfc2_nci_cmd_wq#2008 ->(wq_completion)nfc2_nci_cmd_wq#2009 ->(wq_completion)nfc2_nci_cmd_wq#2010 ->(wq_completion)nfc3_nci_cmd_wq#1723 ->(wq_completion)nfc4_nci_cmd_wq#537 ->(wq_completion)nfc2_nci_cmd_wq#2011 ->(wq_completion)nfc3_nci_cmd_wq#1724 ->(wq_completion)nfc2_nci_cmd_wq#2012 ->(wq_completion)nfc4_nci_cmd_wq#538 ->(wq_completion)nfc3_nci_cmd_wq#1725 ->(wq_completion)nfc5_nci_cmd_wq#367 ->(wq_completion)nfc2_nci_cmd_wq#2013 ->(wq_completion)nfc3_nci_cmd_wq#1726 ->(wq_completion)nfc2_nci_cmd_wq#2014 ->(wq_completion)nfc2_nci_cmd_wq#2015 ->(wq_completion)nfc3_nci_cmd_wq#1727 ->(wq_completion)nfc4_nci_cmd_wq#539 ->(wq_completion)nfc2_nci_cmd_wq#2016 ->(wq_completion)nfc3_nci_cmd_wq#1728 ->(wq_completion)nfc4_nci_cmd_wq#540 ->(wq_completion)nfc11_nci_cmd_wq#13 ->(wq_completion)nfc9_nci_cmd_wq#25 ->(wq_completion)nfc10_nci_cmd_wq#13 ->(wq_completion)nfc7_nci_cmd_wq#82 ->(wq_completion)nfc8_nci_cmd_wq#40 ->(wq_completion)nfc5_nci_cmd_wq#368 ->(wq_completion)nfc6_nci_cmd_wq#120 ->(wq_completion)nfc4_nci_cmd_wq#541 ->(wq_completion)nfc3_nci_cmd_wq#1729 ->(wq_completion)nfc2_nci_cmd_wq#2017 ->(wq_completion)nfc2_nci_cmd_wq#2018 ->(wq_completion)nfc3_nci_cmd_wq#1730 ->(wq_completion)nfc4_nci_cmd_wq#542 ->(wq_completion)nfc5_nci_cmd_wq#369 ->(wq_completion)nfc6_nci_cmd_wq#121 ->(wq_completion)nfc2_nci_cmd_wq#2019 ->(wq_completion)nfc4_nci_cmd_wq#543 ->(wq_completion)nfc8_nci_cmd_wq#41 ->(wq_completion)nfc7_nci_cmd_wq#83 ->(wq_completion)nfc3_nci_cmd_wq#1731 ->(wq_completion)nfc2_nci_cmd_wq#2021 ->(wq_completion)nfc4_nci_cmd_wq#544 ->(wq_completion)nfc2_nci_cmd_wq#2022 ->(wq_completion)nfc3_nci_cmd_wq#1732 ->(wq_completion)nfc4_nci_cmd_wq#545 ->(wq_completion)nfc6_nci_cmd_wq#122 ->(wq_completion)nfc5_nci_cmd_wq#372 ->(wq_completion)nfc2_nci_cmd_wq#2023 ->(wq_completion)nfc2_nci_cmd_wq#2024 ->(wq_completion)nfc3_nci_cmd_wq#1733 FD: 1 BD: 46 +...: &this->info_list_lock FD: 1 BD: 2 ....: (&ndev->cmd_timer) FD: 1 BD: 2 ....: (&ndev->data_timer) FD: 10 BD: 217 +...: prog_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 33 BD: 46 +.-.: (&peer->timer_persistent_keepalive) ->pool_lock#2 ->&list->lock#14 ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->init_task.mems_allowed_seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 24 BD: 5 +.+.: (work_completion)(&rfkill->uevent_work) ->&rq->__lock FD: 1 BD: 1 +.+.: &local->sockets.lock FD: 1 BD: 1 +.+.: &local->raw_sockets.lock FD: 1 BD: 1 ....: (&local->link_timer) FD: 24 BD: 1 +.+.: (work_completion)(&local->tx_work) ->&rq->__lock FD: 24 BD: 1 +.+.: (work_completion)(&local->rx_work) ->&rq->__lock FD: 24 BD: 1 +.+.: (work_completion)(&local->timeout_work) ->&rq->__lock FD: 1 BD: 1 ....: (&local->sdreq_timer) FD: 24 BD: 1 +.+.: (work_completion)(&local->sdreq_timeout_work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1999 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1585 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#95 FD: 25 BD: 8 ....: &sk->sk_lock.wq ->&p->pi_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#525 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1923 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#80 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#523 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#301 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1960 ->&rq->__lock FD: 1 BD: 3 +...: vsock_table_lock FD: 165 BD: 2 +.+.: sk_lock-AF_VSOCK ->slock-AF_VSOCK ->&mm->mmap_lock ->vsock_table_lock ->clock-AF_VSOCK ->&rq->__lock FD: 1 BD: 3 +...: slock-AF_VSOCK FD: 1 BD: 3 +...: clock-AF_VSOCK FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1584 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1534 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1584 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1586 FD: 27 BD: 2 +.+.: sk_lock-AF_NETROM ->&rq->__lock ->slock-AF_NETROM ->&obj_hash[i].lock ->nr_list_lock ->rlock-AF_NETROM FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1436 FD: 1 BD: 2 +.+.: misc_mtx.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1538 FD: 1 BD: 55 .+.-: &table->lock#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1431 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#302 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#301 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1588 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#534 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1838 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#301 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1638 ->&rq->__lock FD: 1 BD: 4 +.+.: genl_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#300 FD: 30 BD: 1 ..-.: &(&conn->info_timer)->timer FD: 2 BD: 52 +.+.: (work_completion)(&(&conn->info_timer)->work) ->&conn->chan_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1839 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#300 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1508 FD: 26 BD: 2 +.+.: sk_lock-AF_KCM ->slock-AF_KCM ->clock-AF_KCM ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +...: clock-AF_PHONET FD: 1 BD: 1 +.+.: (work_completion)(&data->fib_flush_work) FD: 24 BD: 1 +.+.: ®ion->snapshot_lock ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1409 ->&rq->__lock FD: 25 BD: 45 +.+.: __ip_vs_mutex ->&ipvs->dest_trash_lock ->&rq->__lock FD: 1 BD: 46 +...: &ipvs->dest_trash_lock FD: 39 BD: 48 +...: _xmit_NETROM ->(console_sem).lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1384 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1246 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1401 FD: 24 BD: 2 +.+.: &knet->mutex ->&rq->__lock FD: 1 BD: 2 +...: &mux->lock FD: 2 BD: 2 +...: &mux->rx_lock ->rlock-AF_KCM FD: 1 BD: 1 ....: sync_timeline_list_lock FD: 4 BD: 1 ....: &obj->lock ->tk_core.seq.seqcount FD: 1 BD: 1 ....: driver_id_numbers.xa_lock FD: 39 BD: 2 ..-.: &dev->lock ->(console_sem).lock ->console_owner_lock ->console_owner FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1475 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 49 BD: 1 +.+.: &fc->uapi_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&rq->__lock ->&n->list_lock FD: 1 BD: 3 +...: slock-AF_KCM FD: 1 BD: 3 +...: clock-AF_KCM FD: 1 BD: 2 +.+.: (work_completion)(&kcm->tx_work) FD: 1 BD: 3 ....: rlock-AF_KCM FD: 33 BD: 2 +.+.: sk_lock-AF_PHONET ->slock-AF_PHONET ->&pnsocks.lock ->resource_mutex ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->pool_lock ->key ->pcpu_lock ->percpu_counters_lock ->stock_lock FD: 1 BD: 3 +...: slock-AF_PHONET FD: 1 BD: 3 +.+.: &pnsocks.lock FD: 24 BD: 3 +.+.: resource_mutex ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#258 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1603 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1581 ->&rq->__lock FD: 1 BD: 1 +.+.: &resv_map->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1571 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1583 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1920 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1500 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci2#14 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 71 BD: 1 +.+.: &type->s_umount_key#47 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rename_lock.seqcount ->&dentry->d_lock ->&dentry->d_lock/1 ->&sb->s_type->i_lock_key#32 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&rq->__lock ->binderfs_minors_mutex ->rcu_node_0 FD: 1 BD: 5 +.+.: nfc_devlist_mutex.wait_lock FD: 61 BD: 3 +.+.: &type->s_umount_key#48 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#7 ->&rq->__lock FD: 24 BD: 1 +.+.: (work_completion)(&data->suspend_work) ->&rq->__lock FD: 1 BD: 124 +.+.: dev_pm_qos_sysfs_mtx.wait_lock FD: 1 BD: 1 +.+.: (work_completion)(&hdev->reenable_adv_work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#41 FD: 24 BD: 2 +.+.: (wq_completion)nfc15_nci_tx_wq#3 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 6 BD: 1 +.+.: put_task_map-wait-type-override#2 ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_tx_wq#1659 ->&rq->__lock FD: 1 BD: 2 +...: clock-AF_ROSE FD: 24 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq ->&rq->__lock FD: 1 BD: 90 +.+.: freezer_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#361 FD: 30 BD: 1 +.-.: (&sdp->delay_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1961 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1279 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#520 FD: 84 BD: 1 +.-.: (&ndev->rs_timer) ->&ndev->lock ->pool_lock#2 ->&dir->lock#2 ->&ul->lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&zone->lock ->init_task.mems_allowed_seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2010 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1682 ->&rq->__lock FD: 1 BD: 18 +.+.: udc_lock.wait_lock FD: 53 BD: 1 +.-.: (&dum_hcd->timer) ->&dum_hcd->dum->lock ->&dev->lock ->&queue->lock ->&x->wait#27 ->lock#6 ->&x->wait#19 FD: 1 BD: 1 +.+.: (work_completion)(&udc->vbus_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1273 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1492 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1957 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1253 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1262 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1261 FD: 1 BD: 91 +.+.: wq_pool_mutex.wait_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1401 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1439 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1434 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1430 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc52_nci_tx_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1470 ->&rq->__lock FD: 1 BD: 45 +...: &net->xfrm.xfrm_state_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1492 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1337 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1358 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1616 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1373 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1517 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1510 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#471 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#330 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#329 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#367 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1536 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1586 ->&rq->__lock FD: 30 BD: 1 ..-.: &(&bat_priv->tt.work)->timer FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1578 FD: 1420 BD: 2 +.+.: net_cleanup_work ->pernet_ops_rwsem ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->quarantine_lock ->stock_lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 1421 BD: 1 +.+.: (wq_completion)netns ->net_cleanup_work FD: 33 BD: 6 +.+.: (work_completion)(&(&bat_priv->tt.work)->work) ->key#15 ->key#20 ->&bat_priv->tt.req_list_lock ->&bat_priv->tt.roam_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1400 FD: 1 BD: 7 +...: key#20 FD: 1 BD: 7 +...: &bat_priv->tt.req_list_lock FD: 1 BD: 7 +...: &bat_priv->tt.roam_list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1552 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1401 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1553 FD: 1 BD: 5 +...: &tn->node_list_lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#46 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#369 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1556 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1407 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1556 ->&rq->__lock FD: 34 BD: 2 +.+.: (work_completion)(&(&devlink->rwork)->work) ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock#2 ->&cfs_rq->removed.lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 5 ....: (&rxnet->service_conn_reap_timer) FD: 158 BD: 26 +.+.: (wq_completion)hci1#12 ->&rq->__lock ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1482 FD: 1 BD: 126 +.+.: device_links_lock.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1691 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#281 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1509 FD: 1 BD: 47 +...: _xmit_SIT#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1724 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#269 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#270 ->&rq->__lock FD: 1 BD: 47 +...: _xmit_TUNNEL#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1728 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#56 FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#4 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#54 ->&rq->__lock FD: 1 BD: 1 +.+.: &mq_lock FD: 78 BD: 2 +.+.: free_ipc_work ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->mount_lock ->&fsnotify_mark_srcu ->&dentry->d_lock ->&type->s_umount_key#48 ->sb_lock ->unnamed_dev_ida.xa_lock ->list_lrus_mutex ->&xa->xa_lock#5 ->pool_lock#2 ->mnt_id_ida.xa_lock ->&ids->rwsem ->(work_completion)(&ht->run_work) ->&ht->mutex ->percpu_counters_lock ->pcpu_lock ->sysctl_lock ->&sb->s_type->i_lock_key#23 ->rename_lock.seqcount ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->proc_inum_ida.xa_lock ->quarantine_lock ->stock_lock ->&cfs_rq->removed.lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 3 +.+.: &ids->rwsem FD: 1 BD: 47 +...: _xmit_IPGRE#2 FD: 1 BD: 5 +.+.: &fn->fou_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1757 FD: 1 BD: 5 +.+.: ipvs->sync_mutex FD: 1 BD: 49 +.+.: (work_completion)(&sta->drv_deliver_wk) FD: 1 BD: 26 +.+.: (work_completion)(&(&hdev->interleave_scan)->work) FD: 1 BD: 51 +.+.: (work_completion)(&(&conn->id_addr_timer)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1830 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1757 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#298 FD: 1 BD: 5 +.+.: (work_completion)(&local->restart_work) FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#27 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 5 ....: (&local->sta_cleanup) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1783 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1776 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1565 ->&rq->__lock FD: 25 BD: 5 +.+.: rdma_nets_rwsem ->rdma_nets.xa_lock ->&rq->__lock FD: 1 BD: 5 +...: k-clock-AF_NETLINK FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#513 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#68 ->&rq->__lock FD: 1 BD: 45 +.+.: (work_completion)(&(&priv->scan_result)->work) FD: 55 BD: 50 +.+.: (work_completion)(&(&conn->disc_work)->work) ->&hdev->unregister_lock FD: 1 BD: 50 +.+.: (work_completion)(&(&conn->auto_accept_work)->work) FD: 1 BD: 50 +.+.: (work_completion)(&(&conn->idle_work)->work) FD: 1 BD: 79 ....: key#21 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#76 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#2 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#83 FD: 1 BD: 51 +.+.: hci_cb_list_lock.wait_lock FD: 24 BD: 45 +.+.: flowtable_lock ->&rq->__lock FD: 28 BD: 7 +.+.: nf_conntrack_mutex ->&nf_conntrack_locks[i] ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount#7 ->&nf_conntrack_locks[i]/1 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#32 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#497 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#303 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1843 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1959 FD: 24 BD: 45 +.+.: (work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1955 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1619 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1624 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#495 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1651 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#72 FD: 10 BD: 49 +...: &idev->mc_query_lock ->&obj_hash[i].lock FD: 24 BD: 49 +.+.: (work_completion)(&(&idev->mc_report_work)->work) ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#101 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#335 ->&rq->__lock FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#36 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#34 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 45 +.+.: (work_completion)(&port->wq) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1633 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1639 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1914 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#117 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#141 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#481 ->&rq->__lock FD: 24 BD: 49 +.+.: &net->xdp.lock ->&rq->__lock FD: 1 BD: 49 +.+.: mirred_list_lock FD: 1 BD: 49 +...: &idev->mc_report_lock FD: 25 BD: 49 +.+.: &pnn->pndevs.lock ->&rq->__lock FD: 24 BD: 49 +.+.: &pnn->routes.lock ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#35 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1950 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#494 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1939 ->&rq->__lock FD: 1 BD: 45 +...: _xmit_NETROM#2 FD: 24 BD: 1 +.+.: (work_completion)(&(&team->mcast_rejoin.dw)->work) ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1649 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1964 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#5 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1941 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#514 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#5 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#503 ->&rq->__lock FD: 28 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1961 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1422 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1423 FD: 1 BD: 163 +.+.: pcpu_alloc_mutex.wait_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1689 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#110 FD: 41 BD: 5 +.+.: (wq_completion)bond0#20 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 5 +...: &bat_priv->gw.list_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1700 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#349 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1420 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2022 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#539 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#343 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1964 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#346 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1682 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1961 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1680 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1998 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1999 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1988 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1966 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1699 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1696 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1686 ->&rq->__lock FD: 1 BD: 5 +.+.: quirk_mutex FD: 168 BD: 3 +.+.: register_mutex#6 ->fs_reclaim ->pool_lock#2 ->snd_card_mutex ->&x->wait#9 ->&obj_hash[i].lock ->&entry->access ->info_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->mmu_notifier_invalidate_range_start ->&c->lock ->&dum_hcd->dum->lock ->&x->wait#19 ->&base->lock ->&rq->__lock ->(&timer.timer) ->pool_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&____s->seqcount ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&k->k_lock ->subsys mutex#71 ->sound_mutex ->&card->controls_rwsem ->proc_subdir_lock ->proc_inum_ida.xa_lock ->sound_oss_mutex ->strings ->&n->list_lock ->&card->files_lock ->sound_loader_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->req_lock ->&p->pi_lock ->&x->wait#11 ->deferred_probe_mutex ->device_links_lock ->&card->ctl_files_rwlock ->&ent->pde_unload_lock ->&card->power_sleep ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&sem->wait_lock ->&____s->seqcount#2 ->hcd->bandwidth_mutex ->quarantine_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1414 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#512 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1413 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1412 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1371 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#516 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1411 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1364 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1241 ->&rq->__lock FD: 24 BD: 5 +.+.: hashlimit_mutex ->&rq->__lock FD: 1 BD: 5 +.+.: (work_completion)(&(&cnet->ecache.dwork)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1244 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1383 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1469 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1324 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1247 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1392 FD: 1 BD: 3 +.+.: subsys mutex#77 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1378 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1372 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1373 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1377 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1404 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1260 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1260 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1399 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1398 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1464 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1386 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1444 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1281 ->&rq->__lock FD: 1 BD: 4 ....: &uhid->waitq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1397 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1447 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1301 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1446 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1434 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1445 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1433 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1430 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1436 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1435 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#254 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1288 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1432 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1457 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1462 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1461 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1465 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1627 ->&rq->__lock FD: 28 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1628 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock FD: 1 BD: 45 +...: &net->xfrm.xfrm_policy_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1628 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 51 BD: 1 +.-.: (&p->forward_delay_timer) ->&br->lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1622 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1469 ->&rq->__lock FD: 1 BD: 93 +.+.: wq_pool_attach_mutex.wait_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1467 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1491 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1479 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1338 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1346 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1485 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1477 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1482 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1504 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1504 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1618 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1616 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1497 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1619 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1500 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1501 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1503 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1506 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1460 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1615 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#373 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#474 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1515 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1373 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1369 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1514 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1513 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1519 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1595 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1598 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1602 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#365 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1526 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1534 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1528 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1522 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#260 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1563 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1561 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1533 ->&rq->__lock FD: 1 BD: 57 +.-.: &r->producer_lock#3 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1647 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1386 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1378 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1533 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1584 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1433 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1587 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1587 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1537 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1434 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1571 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1545 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1391 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1577 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#80 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1580 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1426 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1579 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#370 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1391 FD: 25 BD: 19 ....: &root->deactivate_waitq ->&p->pi_lock FD: 1 BD: 5 +...: &net->nsid_lock FD: 1 BD: 5 ....: netdev_unregistering_wq.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1539 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1547 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1546 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1651 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1540 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1388 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1539 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1419 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1566 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1544 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#371 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#369 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1559 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1553 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1442 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1549 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1548 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1405 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1439 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#371 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1411 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1408 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#369 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1419 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1433 FD: 1 BD: 47 +...: _xmit_NONE#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1582 FD: 41 BD: 5 +.+.: (wq_completion)bond0#24 ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1431 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1432 ->&rq->__lock FD: 55 BD: 46 +.-.: (&peer->timer_retransmit_handshake) ->&peer->endpoint_lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#47 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1413 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1554 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1561 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1573 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1560 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1573 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1569 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1604 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1600 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1599 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1629 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1627 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1624 ->&rq->__lock FD: 1 BD: 6 +.+.: ebt_mutex.wait_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1622 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1628 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1625 ->&rq->__lock FD: 1 BD: 5 +.+.: netns_bpf_mutex FD: 184 BD: 1 +.+.: (wq_completion)hci5#3 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 5 ....: (&net->fs_probe_timer) FD: 1 BD: 7 ++++: &net->cells_lock FD: 1 BD: 5 ....: (&net->cells_timer) FD: 30 BD: 1 +.+.: (wq_completion)afs ->(work_completion)(&net->cells_manager) ->(work_completion)(&net->fs_manager) FD: 27 BD: 2 +.+.: (work_completion)(&net->cells_manager) ->&net->cells_lock ->bit_wait_table + i ->&rq->__lock FD: 1 BD: 5 ....: (&net->fs_timer) FD: 27 BD: 2 +.+.: (work_completion)(&net->fs_manager) ->&(&net->fs_lock)->lock ->bit_wait_table + i ->&rq->__lock FD: 1 BD: 3 +.+.: &(&net->fs_lock)->lock FD: 1 BD: 6 +.+.: &rx->incoming_lock FD: 1 BD: 6 +.+.: &call->notify_lock FD: 1 BD: 6 ....: (rxrpc_call_limiter).lock FD: 1 BD: 6 +.+.: &rx->recvmsg_lock FD: 1 BD: 6 ....: (&call->timer) FD: 1 BD: 6 ....: &list->lock#17 FD: 1 BD: 5 +.+.: (wq_completion)kafsd FD: 1 BD: 5 +...: k-clock-AF_RXRPC FD: 1 BD: 5 ..-.: rlock-AF_RXRPC FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1633 FD: 1 BD: 1 ....: (&local->client_conn_reap_timer) FD: 1 BD: 1 ....: &list->lock#18 FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#2 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1480 ->&rq->__lock FD: 1 BD: 8 +.+.: (work_completion)(&data->gc_work) FD: 1 BD: 5 +.+.: (work_completion)(&ovs_net->dp_notify_work) FD: 1 BD: 5 +...: &srv->idr_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc56_nci_tx_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1646 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq ->&rq->__lock FD: 1 BD: 7 +...: &nt->cluster_scope_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy49 FD: 1 BD: 5 +.+.: (work_completion)(&(&c->work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq FD: 1 BD: 5 +.+.: (wq_completion)krdsd FD: 1 BD: 5 +.+.: (work_completion)(&rtn->rds_tcp_accept_w) FD: 1 BD: 49 +...: &icsk->icsk_accept_queue.rskq_lock#2 FD: 1 BD: 5 ....: rds_tcp_conn_lock FD: 1 BD: 5 ....: loop_conns_lock FD: 1 BD: 5 +.+.: (wq_completion)l2tp FD: 2 BD: 6 +.+.: (work_completion)(&rxnet->service_conn_reaper) ->&rxnet->conn_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1662 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1665 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1662 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1667 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1670 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#36 ->&rq->__lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#27 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1522 FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc47_nci_tx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1686 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#282 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#20 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1696 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#380 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1732 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1514 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#399 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#59 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#271 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1709 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#39 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#38 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#383 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1720 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#266 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1498 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#50 FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#6 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1502 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#6 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1723 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1725 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1724 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#54 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#387 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#53 FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#3 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#53 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#3 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#42 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#6 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1502 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#387 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#41 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#270 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#42 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1783 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#265 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#280 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#50 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#383 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#58 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#6 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#61 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#11 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#2 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#2 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#5 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1725 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#14 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#4 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#56 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1504 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1511 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1509 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1511 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#56 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1731 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#6 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#392 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#389 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1731 ->&rq->__lock FD: 29 BD: 1 +.-.: (&timer) ->&obj_hash[i].lock ->&base->lock ->&txlock ->&txwq FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#12 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#278 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#64 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#123 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1741 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#408 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1521 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#70 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#23 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1537 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#114 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#116 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#101 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1570 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1822 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1818 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1572 FD: 41 BD: 5 +.+.: (wq_completion)bond0#28 ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#416 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1640 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#109 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#434 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#295 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1825 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#79 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#437 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#437 FD: 1 BD: 49 ....: (&ifibss->timer) FD: 850 BD: 6 +.+.: (work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 45 +.+.: (work_completion)(&wdev->disconnect_wk) FD: 1 BD: 45 +.+.: (work_completion)(&wdev->pmsr_free_wk) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#436 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1537 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1537 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1536 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1538 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1581 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1535 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#287 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#72 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1752 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1751 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1754 FD: 1 BD: 49 ....: (&local->dynamic_ps_timer) FD: 1 BD: 49 ....: (&dwork->timer)#3 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#53 FD: 1 BD: 45 ....: &rdev->dev_wait FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#286 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1751 FD: 1 BD: 49 ....: (&dwork->timer)#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1822 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1572 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1821 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1578 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1820 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#105 FD: 1 BD: 2 +.+.: nf_sockopt_mutex.wait_lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#105 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1570 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#92 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#55 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#420 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#28 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#29 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1765 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#419 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#45 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#40 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#58 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#43 FD: 1 BD: 5 +.+.: (work_completion)(&rdev->conn_work) FD: 1 BD: 5 +.+.: (work_completion)(&(&rdev->background_cac_done_wk)->work) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->destroy_work) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->propagate_radar_detect_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->propagate_cac_done_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->mgmt_registrations_update_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->background_cac_abort_wk) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1287 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#52 FD: 158 BD: 26 +.+.: (wq_completion)hci1#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 184 BD: 1 +.+.: (wq_completion)hci1#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#51 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#51 FD: 30 BD: 1 ..-.: &(&bat_priv->dat.work)->timer FD: 30 BD: 1 ..-.: &(&bat_priv->bla.work)->timer FD: 25 BD: 6 +.+.: (work_completion)(&(&bat_priv->dat.work)->work) ->&hash->list_locks[i] ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 7 +...: &hash->list_locks[i] FD: 32 BD: 6 +.+.: (work_completion)(&(&bat_priv->bla.work)->work) ->&rq->__lock ->key#19 ->&obj_hash[i].lock ->&base->lock ->crngs.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1656 FD: 184 BD: 1 +.+.: (wq_completion)hci2#13 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#48 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1782 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1776 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#30 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1812 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1558 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#29 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#57 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1790 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1782 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1549 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#293 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#428 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1799 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1930 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#431 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#431 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#87 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1566 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy42 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1563 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1555 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1794 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#74 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1796 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#427 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1555 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1800 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#423 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1793 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#65 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#65 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#80 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#60 FD: 1 BD: 93 +.+.: gdp_mutex.wait_lock FD: 1 BD: 45 +.+.: (work_completion)(&(&priv->connect)->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#72 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#74 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1684 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#82 FD: 59 BD: 1 .+.+: sb_writers#12 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#27 ->&wb->list_lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#62 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#61 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1794 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1802 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#324 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1562 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1563 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy41 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#299 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#97 FD: 43 BD: 46 +.+.: (work_completion)(&br->mcast_gc_work) ->&br->multicast_lock ->(&p->rexmit_timer) ->&obj_hash[i].lock ->&base->lock ->(&p->timer) ->pool_lock#2 ->krc.lock ->(&mp->timer) ->&rq->__lock FD: 1 BD: 47 ....: (&p->rexmit_timer) FD: 1 BD: 47 ....: (&p->timer) FD: 39 BD: 47 +.-.: (&mp->timer) ->&br->multicast_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#429 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1802 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1804 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#58 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#57 FD: 1 BD: 45 ....: (&pmctx->ip6_mc_router_timer) FD: 1 BD: 45 ....: (&pmctx->ip4_mc_router_timer) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1838 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1837 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#312 FD: 850 BD: 2 +.+.: (crda_timeout).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 30 BD: 1 ..-.: net/wireless/reg.c:533 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#453 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1852 FD: 1 BD: 47 +...: &qdisc_xmit_lock_key FD: 1 BD: 47 +...: &qdisc_xmit_lock_key#2 FD: 1 BD: 47 +...: &vlan_netdev_xmit_lock_key FD: 1 BD: 47 +...: &batadv_netdev_xmit_lock_key FD: 1 BD: 47 +...: &qdisc_xmit_lock_key#3 FD: 1 BD: 47 +...: &qdisc_xmit_lock_key#4 FD: 1 BD: 47 +...: _xmit_LOOPBACK#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1846 FD: 26 BD: 1 +.+.: put_task_map-wait-type-override#3 ->&rq->__lock ->&obj_hash[i].lock ->&____s->seqcount ->pool_lock#2 ->stock_lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#115 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#119 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#133 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1865 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1872 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1872 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#105 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#464 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#338 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1674 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1642 ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#62 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&cfs_rq->removed.lock FD: 24 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#6 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1669 FD: 32 BD: 46 +.-.: (&peer->timer_send_keepalive) ->&c->lock ->pool_lock#2 ->&list->lock#14 ->tk_core.seq.seqcount ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 46 ....: (&peer->timer_new_handshake) FD: 1 BD: 46 ....: (&peer->timer_zero_key_material) FD: 1 BD: 46 +.+.: (work_completion)(&peer->clear_peer_work) FD: 1 BD: 45 +.+.: (work_completion)(&(&bond->arp_work)->work) FD: 1 BD: 45 +.+.: (work_completion)(&(&bond->alb_work)->work) FD: 1 BD: 45 +.+.: (work_completion)(&(&bond->ad_work)->work) FD: 1 BD: 45 +.+.: (work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 45 +.+.: (work_completion)(&(&bond->slave_arr_work)->work) FD: 1 BD: 45 ....: (&br->hello_timer) FD: 1 BD: 45 ....: (&br->topology_change_timer) FD: 1 BD: 45 ....: (&br->tcn_timer) FD: 1 BD: 45 ....: (&brmctx->ip4_mc_router_timer) FD: 1 BD: 45 ....: (&brmctx->ip4_other_query.timer) FD: 1 BD: 45 ....: (&brmctx->ip4_other_query.delay_timer) FD: 1 BD: 45 ....: (&brmctx->ip6_mc_router_timer) FD: 1 BD: 45 ....: (&brmctx->ip6_other_query.timer) FD: 1 BD: 45 ....: (&brmctx->ip6_other_query.delay_timer) FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#342 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#36 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1619 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1617 ->&rq->__lock FD: 41 BD: 1 +.+.: (wq_completion)bond0#33 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#32 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1624 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1902 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1958 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1621 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1666 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#494 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#101 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#482 FD: 158 BD: 1 +.+.: (wq_completion)hci3#10 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#103 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#72 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1650 ->&rq->__lock FD: 1 BD: 1 +.+.: virtio_9p_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1936 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1945 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1650 ->&rq->__lock FD: 133 BD: 46 +.+.: team->team_lock_key#33 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->&rq->__lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->lweventlist_lock ->(console_sem).lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#32 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#63 ->(work_completion)(&peer->transmit_handshake_work) FD: 184 BD: 1 +.+.: (wq_completion)hci5#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 184 BD: 1 +.+.: (wq_completion)hci1#15 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 1 +.+.: (wq_completion)hci5#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#335 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1948 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#337 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1907 ->&rq->__lock FD: 158 BD: 1 +.+.: (wq_completion)hci2#16 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1667 ->&rq->__lock FD: 1 BD: 45 +.+.: raw_notifier_lock FD: 1 BD: 45 +.+.: bcm_notifier_lock FD: 1 BD: 45 +.+.: isotp_notifier_lock FD: 158 BD: 1 +.+.: (wq_completion)hci1#16 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1664 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#17 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci0#12 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1712 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1647 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1417 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1656 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#331 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#37 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#94 FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg0#68 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1912 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1630 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1629 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#475 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#324 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1911 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1908 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1635 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1634 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1918 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1919 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1915 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1637 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1929 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1925 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#70 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1918 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1641 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1923 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2002 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#524 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#496 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1955 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1942 FD: 1 BD: 3 +...: rose_list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1930 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#4 FD: 24 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#497 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#498 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq FD: 1 BD: 45 +...: &bond->ipsec_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#338 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1670 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#501 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1662 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1687 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#101 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1935 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#336 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1663 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#490 ->&rq->__lock FD: 1 BD: 1 +.+.: (work_completion)(&(&team->notify_peers.dw)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#328 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#502 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1672 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#23 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1962 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#496 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#500 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1952 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#9 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1665 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#10 FD: 1 BD: 45 +...: &pmc->lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1662 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1661 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#488 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#17 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#492 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1939 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1659 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1937 ->&rq->__lock FD: 1 BD: 1 .+.+: drm_unplug_srcu FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#484 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1651 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#482 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1649 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#97 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#485 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1937 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#112 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#352 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1979 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#113 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#513 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#353 ->&rq->__lock FD: 1 BD: 53 ..-.: key#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1974 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1933 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1931 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#96 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#19 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1664 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#19 ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#21 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#363 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1965 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2000 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1991 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#360 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#521 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1709 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1708 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1997 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1426 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1424 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1425 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1280 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#109 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1574 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1279 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#112 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1690 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#350 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1695 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#428 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1695 FD: 32 BD: 1 +.-.: (&cfile->notify_timer) ->cgroup_file_kn_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#78 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#114 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1984 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1980 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1975 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#515 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#355 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1701 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#356 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1979 FD: 1 BD: 5 +.+.: (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) FD: 4 BD: 5 +.+.: &bat_priv->bat_v.ogm_buff_mutex ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#522 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1976 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1705 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1984 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1707 ->&rq->__lock FD: 1 BD: 5 +.+.: &hn->hn_lock FD: 60 BD: 1 +.+.: (wq_completion)inet_frag_wq ->(work_completion)(&fqdir->destroy_work) FD: 59 BD: 2 +.+.: (work_completion)(&fqdir->destroy_work) ->(work_completion)(&ht->run_work) ->&ht->mutex FD: 41 BD: 2 +.+.: fqdir_free_work ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->quarantine_lock FD: 27 BD: 45 +.+.: &caifn->caifdevs.lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock#2 ->&this->info_list_lock ->&cfs_rq->removed.lock ->pool_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1706 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#513 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#514 ->&rq->__lock FD: 1 BD: 5 +.+.: &pnetids_ndev->lock FD: 28 BD: 5 +.+.: k-sk_lock-AF_INET6/1 ->k-slock-AF_INET6 ->rlock-AF_INET6 ->&list->lock#19 FD: 1 BD: 6 ....: rlock-AF_INET6 FD: 1 BD: 8 ....: &list->lock#19 FD: 8 BD: 49 +...: k-slock-AF_INET6/1 ->&sctp_ep_hashtable[i].lock ->&obj_hash[i].lock ->pool_lock#2 ->k-clock-AF_INET6 ->key#24 FD: 1 BD: 51 +...: &sctp_ep_hashtable[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#114 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#518 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#517 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#519 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#516 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#512 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1693 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#78 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#114 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1420 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1421 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#349 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1277 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#112 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1419 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1418 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1277 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1419 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1275 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1418 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1418 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#367 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#369 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#6 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2018 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1968 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#537 FD: 30 BD: 1 ..-.: &(&conn->disc_work)->timer FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1724 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1966 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1962 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1959 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1682 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1679 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#107 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#507 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#344 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1684 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#508 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1683 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1683 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#347 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#109 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#347 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#107 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#507 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#109 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#37 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1957 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1679 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1725 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1967 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1968 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1679 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#509 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1726 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1726 ->&rq->__lock FD: 855 BD: 13 +.+.: &devlink->lock_key#38 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&xa->xa_lock#19 ->&c->lock ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->&(&fn_net->fib_chain)->lock ->stack_depot_init_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&n->list_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1723 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1721 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2006 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1438 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#364 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2006 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#535 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#117 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1714 FD: 133 BD: 46 +.+.: team->team_lock_key#38 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&c->lock ->&n->list_lock ->lweventlist_lock ->(console_sem).lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount ->quarantine_lock FD: 41 BD: 1 +.+.: (wq_completion)bond0#38 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#1713 FD: 25 BD: 31 ..-.: semaphore->lock#2 ->&p->pi_lock FD: 26 BD: 30 ..-.: &queue->lock ->pool_lock#2 ->semaphore->lock#2 ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1529 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1525 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1704 ->&rq->__lock FD: 25 BD: 2 ..-.: &x->wait#27 ->&p->pi_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#515 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1979 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1967 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#350 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1975 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1220 FD: 1 BD: 226 ...-: init_task.mems_allowed_seq.seqcount FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1366 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1221 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1363 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1365 ->&rq->__lock FD: 74 BD: 1 +.-.: (&sk->sk_timer) ->slock-AF_INET#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1367 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1368 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1228 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1413 FD: 1 BD: 4 ....: &card->power_sleep FD: 1 BD: 4 +.+.: &card->files_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1408 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1416 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1417 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#542 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2011 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1724 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2009 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#118 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2007 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1417 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1416 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1274 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#367 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1273 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1273 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1415 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#22 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1370 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1410 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1269 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1412 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1413 FD: 1 BD: 3 +.+.: oom_adj_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1371 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1368 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1227 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1368 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1366 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1367 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1362 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1224 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1361 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1566 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2024 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1226 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1364 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1363 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1230 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1228 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1369 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1493 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1621 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1622 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1623 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#348 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1380 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1379 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1240 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1381 ->&rq->__lock FD: 25 BD: 5 +.+.: &sn->gssp_lock ->rcu_node_0 ->&rq->__lock FD: 1 BD: 8 +.+.: &cd->hash_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1382 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1381 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1243 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1384 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1384 FD: 1 BD: 5 +.+.: xfrm_state_gc_work FD: 1 BD: 5 +...: ip6_fl_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1385 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1265 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1264 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1406 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#353 FD: 1 BD: 5 ....: (&net->ipv6.ip6_fib_timer) FD: 1 BD: 45 ....: (&mrt->ipmr_expire_timer) FD: 30 BD: 1 ..-.: &(&tbl->gc_work)->timer FD: 44 BD: 2 +.+.: (work_completion)(&(&tbl->gc_work)->work) ->&tbl->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 5 ....: (&ipvs->dest_trash_timer) FD: 1 BD: 5 +.+.: (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) FD: 24 BD: 5 +.+.: (work_completion)(&(&ipvs->est_reload_work)->work) ->&rq->__lock FD: 1 BD: 5 +...: recent_lock FD: 1 BD: 5 +.+.: trans_gc_work FD: 1 BD: 52 +.+.: rcu_state.exp_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1408 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1407 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1408 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1409 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1411 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#354 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1407 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1407 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1266 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1405 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1404 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1264 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1385 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1382 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1383 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1381 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1380 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#374 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1472 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1471 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1470 FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.policy_hash_work) FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.state_hash_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1469 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1387 ->&rq->__lock FD: 2 BD: 3 ....: &uhid->qlock ->&uhid->waitq FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1387 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1246 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1388 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#349 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1389 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1388 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1389 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1391 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1392 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1393 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1392 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1394 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1396 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1254 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1397 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1376 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1237 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1377 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#348 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#348 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1397 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1402 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1380 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1237 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1375 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#347 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1234 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1375 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1372 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#247 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1375 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1374 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1376 ->&rq->__lock FD: 157 BD: 2 +.+.: (work_completion)(&uhid->worker) ->dquirks_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&k->list_lock ->&c->lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&n->list_lock ->subsys mutex#77 ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock ->rcu_node_0 ->&____s->seqcount#2 ->&rcu_state.expedited_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1372 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#346 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1373 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1371 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1231 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1231 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1374 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#347 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1242 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1382 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1383 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1238 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1396 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1399 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1398 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1258 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1399 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1259 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1401 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1261 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1403 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1403 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1402 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1402 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1258 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1400 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1400 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1395 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#350 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1396 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1393 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1391 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1390 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1247 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1387 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1388 FD: 26 BD: 4 +.+.: subsys mutex#78 ->&rq->__lock ->&k->k_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1466 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1442 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1441 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1443 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1446 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1445 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1302 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1424 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1412 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1270 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1391 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#250 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1390 FD: 126 BD: 3 +.+.: minors_rwsem ->fs_reclaim ->&c->lock ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->&k->k_lock ->subsys mutex#78 ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&n->list_lock ->&sem->wait_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1394 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1400 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1398 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#351 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1266 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1410 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1403 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1419 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1417 FD: 1 BD: 48 +.+.: rcu_state.barrier_mutex.wait_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1416 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1420 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1303 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1426 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1445 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1442 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1443 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1444 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1300 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1317 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc50_nci_cmd_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc51_nci_tx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc51_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1306 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1450 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1448 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1303 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1442 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1431 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1431 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1437 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1440 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1298 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1297 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1441 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1440 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1298 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#43 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#43 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#44 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1434 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1436 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1431 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1432 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1433 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1297 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1441 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1429 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1289 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1288 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1436 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1432 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1444 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1443 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1457 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1310 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1454 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#358 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1307 FD: 1 BD: 2 +.+.: (wq_completion)nfc51_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc52_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1461 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1573 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1462 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1464 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1464 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1465 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1323 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1467 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1467 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1466 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1468 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1470 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1471 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1329 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1473 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1472 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1626 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1473 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#49 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#50 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc50_nci_tx_wq FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#49 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc49_nci_cmd_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1625 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc50_nci_rx_wq FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#50 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#49 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1472 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1327 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1328 ->&rq->__lock FD: 15 BD: 1 +.-.: (&n->timer) ->&n->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1468 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1466 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1465 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1461 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1462 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1460 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1459 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1460 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1463 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1463 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1326 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#374 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1472 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1620 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1468 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1478 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1477 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1337 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1481 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1483 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1482 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1485 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1484 ->&rq->__lock FD: 35 BD: 2 +.+.: (work_completion)(&pool->idle_cull_work) ->wq_pool_attach_mutex ->wq_pool_attach_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1486 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1487 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1489 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1488 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1491 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1490 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1348 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1490 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1345 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1490 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1486 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1341 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1480 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1485 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1481 FD: 1 BD: 1 +...: btf_idr_lock FD: 30 BD: 1 +.-.: (&pool->idle_timer) ->&pool->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1478 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1480 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1476 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1479 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1475 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1475 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1474 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#359 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1473 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1472 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1481 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1480 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1487 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1344 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1488 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1491 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1489 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1359 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1506 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1509 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1509 FD: 1 BD: 3 +.+.: tty_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1618 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1617 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1465 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1617 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1617 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1506 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1505 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1362 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1508 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1503 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1505 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1501 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1502 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1358 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1353 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#361 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1495 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1496 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1508 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1507 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1504 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1500 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1500 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1499 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1497 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1494 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1494 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1357 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1505 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1507 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1363 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1362 FD: 1 BD: 49 +.+.: (wq_completion)phy48 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1615 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1614 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1462 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1616 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1615 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1462 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1612 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1459 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1608 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1609 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1610 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1456 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1605 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1606 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1455 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1522 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#364 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1519 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#363 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1511 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1512 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1512 ->&rq->__lock FD: 5 BD: 2 +.+.: (ima_keys_delayed_work).work ->ima_keys_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1510 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1509 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1367 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1518 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1600 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1450 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1603 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1605 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1637 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#473 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#474 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1521 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1605 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1598 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1599 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1594 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1443 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1591 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1532 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1385 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1529 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1383 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#260 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1526 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1524 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1377 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1523 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1523 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1378 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1562 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1563 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1566 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1936 FD: 1 BD: 3 +...: slock-AF_ROSE FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1533 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1529 ->&rq->__lock FD: 10 BD: 218 +...: map_idr_lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->&n->list_lock FD: 6 BD: 2 +.+.: (work_completion)(&map->work) ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1568 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1474 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#23 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1429 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1580 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1580 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1429 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1540 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1538 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1538 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1577 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1578 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1578 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1546 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1545 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#370 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1547 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1425 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1427 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1572 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1541 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1543 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1542 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1543 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1535 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1389 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1390 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1539 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1536 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1536 ->&rq->__lock FD: 1 BD: 3 ....: rlock-AF_NETROM FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#18 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1571 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#18 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#18 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1545 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1418 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1569 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1569 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1723 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1953 ->&rq->__lock FD: 1 BD: 3 +...: slock-AF_NETROM FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1436 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1436 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1435 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#371 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1441 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1437 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1593 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1441 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1441 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy43 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1443 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1551 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1403 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1402 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1548 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1550 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1549 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1399 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1552 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1551 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1549 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1554 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1552 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1440 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1440 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1440 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1406 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1555 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1555 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1410 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1560 ->&rq->__lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#24 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1418 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1419 ->&rq->__lock FD: 1 BD: 47 ..-.: &list->lock#20 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1671 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1575 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#45 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#23 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1434 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1583 ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#45 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc15_nci_rx_wq#3 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#334 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#45 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#370 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1425 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1423 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1423 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1407 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1409 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1562 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1562 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1417 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1417 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#48 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1634 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#48 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#24 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 24 BD: 49 +.+.: (wq_completion)phy45 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1558 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1558 ->&rq->__lock FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg2#71 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1413 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1564 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1414 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1415 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1567 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1427 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1577 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1431 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1581 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1588 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#61 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1442 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1594 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1599 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1453 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#20 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy47 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1627 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1626 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1474 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1626 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1623 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1468 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1621 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1620 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1629 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1655 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1630 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1631 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1477 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1636 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1645 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1642 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1651 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#3 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1659 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#376 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1681 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1674 FD: 24 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1673 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1670 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1663 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1662 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci1#11 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1663 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1661 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#37 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1668 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1661 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1664 FD: 41 BD: 5 +.+.: (wq_completion)bond0#26 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1674 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1675 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1679 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#375 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#22 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1483 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1683 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1738 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1684 FD: 1 BD: 92 +.+.: &pa->pa_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1685 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1692 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1738 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#405 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1737 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#404 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#279 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1516 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#399 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#20 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#282 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#282 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1696 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#283 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#68 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#68 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1517 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#379 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1704 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1708 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1703 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#395 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1510 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#392 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#389 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#381 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1714 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1514 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#385 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#10 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1499 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1499 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#38 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#38 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#265 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#265 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#49 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#49 ->&rq->__lock FD: 1 BD: 1 +...: &wqueue->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1500 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#385 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#54 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1723 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#270 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#54 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#42 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1502 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#53 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#269 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#5 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#388 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#388 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#387 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1725 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1500 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#267 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1718 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1718 ->&rq->__lock FD: 1 BD: 45 +...: nr_neigh_list_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#13 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#5 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#7 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#49 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#397 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#4 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1729 FD: 5 BD: 46 +...: _xmit_SLIP#2 ->&eql->queue.lock FD: 5 BD: 1 +...: &nr_netdev_xmit_lock_key ->nr_node_list_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +...: nr_node_list_lock FD: 5 BD: 1 +...: _xmit_X25#2 ->&lapbeth->up_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1520 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#12 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1521 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1571 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1820 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1823 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1828 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#434 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1579 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#79 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#79 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#413 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#287 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1535 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1753 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#54 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#13 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#54 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#297 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1578 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#93 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1569 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#101 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#95 FD: 12 BD: 1 +.+.: &xa->xa_lock#20 ->stock_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#55 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1769 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#418 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#88 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#45 FD: 41 BD: 5 +.+.: (wq_completion)bond0#29 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 41 BD: 5 +.+.: (wq_completion)bond0#15 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#53 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#52 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#50 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#50 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#44 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#46 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#48 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#426 ->&rq->__lock FD: 30 BD: 1 ..-.: &(&net->ipv6.addr_chk_work)->timer FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1774 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1772 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1790 FD: 1 BD: 2 +.+.: loop_validate_mutex.wait_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1812 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1562 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#294 ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#31 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#31 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#87 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#73 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#76 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#59 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#73 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#71 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#80 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#68 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1552 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#61 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#61 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1924 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#88 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1798 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1794 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1798 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1567 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1804 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1810 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#88 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#328 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#302 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#126 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_rx_wq#25 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#124 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#123 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1595 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#118 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1595 FD: 184 BD: 1 +.+.: (wq_completion)hci3#9 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1855 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#315 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#317 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#461 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1873 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#505 FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#12 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1619 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1621 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1898 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#11 FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg0#69 ->(work_completion)(&peer->transmit_handshake_work) FD: 27 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1904 ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1926 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1934 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#329 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#34 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#495 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#11 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#336 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1905 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1900 FD: 24 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#4 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#487 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1937 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1648 FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg2#67 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#74 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1904 FD: 24 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1909 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#323 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1654 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1917 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#473 ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#58 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1930 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#361 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1898 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#22 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#141 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#496 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#141 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#481 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1934 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1652 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1653 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#332 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#483 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#329 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#68 FD: 1 BD: 3 ....: rlock-AF_ROSE FD: 27 BD: 2 +.+.: sk_lock-AF_ROSE ->slock-AF_ROSE ->rose_list_lock ->&obj_hash[i].lock ->rlock-AF_ROSE ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1665 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1928 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1929 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1687 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#494 FD: 24 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1944 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1933 FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#11 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1940 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#482 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#97 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#528 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1957 FD: 1 BD: 2 +...: clock-AF_NETROM FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#12 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#336 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#34 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#9 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#492 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1658 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#99 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#99 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1935 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#484 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1938 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#484 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1652 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1927 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#481 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#333 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#18 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#19 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1714 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#118 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1988 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#527 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci5#4 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1425 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1423 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1424 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1421 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1422 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#348 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#348 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1973 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1973 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#515 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1698 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#523 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1276 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1731 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2016 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#121 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2014 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#24 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2012 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2015 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1970 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1681 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1963 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1549 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#347 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#107 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1681 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1678 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1678 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1725 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1728 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#534 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy76 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2007 FD: 41 BD: 5 +.+.: (wq_completion)bond0#21 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#147 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2004 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#117 ->&rq->__lock FD: 51 BD: 1 +.+.: &udc->connect_lock ->&dum_hcd->dum->lock ->hcd_root_hub_lock ->&queue->lock ->&rq->__lock ->udc_lock FD: 69 BD: 2 +.+.: (work_completion)(&hcd->wakeup_work) FD: 1 BD: 7 +.+.: usbfs_mutex FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1995 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#525 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#524 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1707 FD: 80 BD: 3 +.+.: &port_dev->status_lock ->&hub->status_mutex ->hcd->address0_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->hcd_root_hub_lock ->fs_reclaim ->&dum_hcd->dum->lock ->&obj_hash[i].lock ->&____s->seqcount ->&rq->__lock ->&x->wait#19 ->&c->lock ->&base->lock ->(&timer.timer) ->&n->list_lock FD: 78 BD: 4 +.+.: hcd->address0_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->devtree_lock ->&dev->power.lock ->device_state_lock ->&bus->devnum_next_mutex ->mmu_notifier_invalidate_range_start ->ehci_cf_port_reset_rwsem ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->&____s->seqcount ->&c->lock ->&dum_hcd->dum->lock ->&x->wait#19 ->&base->lock ->(&timer.timer) ->hcd_urb_list_lock ->&n->list_lock ->quirk_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->hcd_root_hub_lock ->&queue->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount#2 FD: 1 BD: 5 +.+.: &bus->devnum_next_mutex FD: 68 BD: 5 .+.+: ehci_cf_port_reset_rwsem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&rq->__lock ->&c->lock ->hcd_root_hub_lock ->fs_reclaim ->&dum_hcd->dum->lock ->&queue->lock ->&obj_hash[i].lock ->&x->wait#19 ->&base->lock ->(&timer.timer) ->&hub->status_mutex ->device_state_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->pool_lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&n->list_lock FD: 39 BD: 1 +.-.: (&hcd->rh_timer) ->&dum_hcd->dum->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#357 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#353 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1698 ->&rq->__lock FD: 1 BD: 51 ....: key#23 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1985 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#516 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1983 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1693 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1975 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#510 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#512 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1692 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1971 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#351 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#509 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1691 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1684 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1685 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#108 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1688 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1973 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#350 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1987 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#523 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#119 ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#41 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2019 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1361 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1366 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1367 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1369 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1369 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1370 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1411 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1493 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1478 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1476 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1332 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1330 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1473 ->&rq->__lock FD: 28 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1474 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1477 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1476 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1338 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1484 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1484 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1488 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1489 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1347 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1492 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1619 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1620 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1619 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1502 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1503 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1502 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1357 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1356 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1501 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1355 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1354 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#361 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1496 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1495 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1496 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1495 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1350 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1493 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1494 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1464 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1612 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1613 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1611 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1610 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1457 FD: 1 BD: 50 +.+.: stack_depot_init_mutex.wait_lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#372 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1606 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1454 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#474 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1520 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1521 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1520 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1520 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#364 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1516 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1517 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1514 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1369 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1511 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1511 ->&rq->__lock FD: 30 BD: 1 ..-.: security/integrity/ima/ima_queue_keys.c:35 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1512 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1367 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1513 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1368 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1513 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1514 FD: 1 BD: 1 +...: &list->lock#21 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1519 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1597 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1596 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1597 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1592 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1592 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1590 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1438 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1589 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1535 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1534 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1535 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1530 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1531 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1531 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1527 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1527 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1378 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1523 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1524 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1524 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#365 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1525 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1525 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1375 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1521 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1522 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1377 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1378 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1379 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#365 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1561 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1437 FD: 1 BD: 2 +.+.: &iopt->iova_rwsem FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1439 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1439 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci2#10 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1594 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1554 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1550 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1551 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1550 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1547 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1400 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1548 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1401 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1402 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1405 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1442 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci2#9 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1588 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1589 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1438 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1557 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1557 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1556 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1555 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1406 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1407 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1408 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1411 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1411 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1412 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1414 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1416 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1416 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1416 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#47 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#48 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#47 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 49 +.+.: (wq_completion)phy46 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1420 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1420 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1421 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1422 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#19 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1423 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1424 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1424 FD: 1 BD: 1 +.-.: x25_list_lock FD: 1 BD: 1 +.-.: x25_forward_list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1425 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1426 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1427 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1428 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1576 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1428 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1428 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1430 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1430 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1582 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1434 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1435 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1435 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1433 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1432 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1432 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1581 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1676 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#23 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1429 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1430 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#46 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 22 BD: 2 +.+.: (work_completion)(&ns->work) ->sysctl_lock ->&obj_hash[i].lock ->pool_lock#2 ->keyring_name_lock ->proc_inum_ida.xa_lock ->stock_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1575 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1424 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1574 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1422 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1422 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1421 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1421 ->&rq->__lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#24 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1557 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1558 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1559 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1559 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1410 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1410 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1409 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1560 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1412 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1563 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1412 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1564 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1564 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1415 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1415 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1414 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1565 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1565 FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#3 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1567 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1418 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1568 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1568 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1567 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1570 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1570 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1420 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1572 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1570 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1574 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1572 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1576 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1575 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1426 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1576 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1579 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1579 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1582 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1583 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1585 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1587 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1588 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1437 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1590 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1590 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1593 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1514 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1445 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1445 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1447 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1597 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1600 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1601 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1601 FD: 184 BD: 1 +.+.: (wq_completion)hci4#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1601 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1602 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1602 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1604 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1604 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1451 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#471 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#471 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#473 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1606 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#5 FD: 41 BD: 5 +.+.: (wq_completion)bond0#25 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1610 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#20 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1609 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1611 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1795 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1614 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1614 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1613 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1463 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1618 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1621 FD: 1 BD: 2 +.+.: (wq_completion)nfc52_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1480 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1727 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1656 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1657 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1479 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1654 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#1631 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1650 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1631 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1632 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1632 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1477 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1476 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1476 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1636 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1637 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1637 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1637 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1639 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1640 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1645 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1646 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1643 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1644 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1654 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1658 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1657 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1658 FD: 1 BD: 2 +.+.: (wq_completion)nfc55_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc55_nci_rx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc55_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#2 FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#2 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#2 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1481 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1482 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#37 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1481 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#37 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#376 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#376 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#264 FD: 24 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1681 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1484 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1674 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1675 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1482 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1483 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1672 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1673 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1671 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1672 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1672 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1671 FD: 24 BD: 49 +.+.: (wq_completion)phy50 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1878 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq FD: 41 BD: 5 +.+.: (wq_completion)bond0#27 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1669 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#96 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#21 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1669 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1669 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1665 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1666 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1664 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1665 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1663 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1660 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1661 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1664 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1666 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1667 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1668 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1666 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1667 FD: 184 BD: 1 +.+.: (wq_completion)hci2#11 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 26 +.+.: (wq_completion)hci2#12 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#21 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1668 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1670 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1676 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1678 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#54 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#27 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#27 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#36 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#22 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#22 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#375 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#375 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1680 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1486 FD: 1 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc41_nci_cmd_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#407 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1683 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1682 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#407 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1739 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#406 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#406 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#284 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1685 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1689 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1686 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1687 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1692 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1687 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1688 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1690 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1689 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1693 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1693 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1692 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1695 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1694 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1694 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1694 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1693 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1737 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1737 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1736 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1736 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#402 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#11 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#278 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1697 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#277 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#19 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#277 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#50 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#63 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#396 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1491 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#379 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1491 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1703 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#378 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1487 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#1488 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1488 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1708 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1709 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1709 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#10 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1511 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#395 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#61 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1509 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#48 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#59 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#274 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#273 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1506 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1710 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#58 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#6 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#58 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#57 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1726 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#6 FD: 24 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#2 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1712 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1712 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1495 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1713 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1716 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1504 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#1716 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1503 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1503 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#55 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#55 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1724 FD: 1 BD: 45 +...: &bpq_netdev_addr_lock_key FD: 1 BD: 46 +...: nr_list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#12 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#386 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#386 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1780 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1722 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#268 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#269 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#3 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#12 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#62 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1730 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#394 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#64 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#398 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#398 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1787 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1779 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#400 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#400 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1518 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#401 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#403 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#403 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#283 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1738 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#405 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#408 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1523 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1743 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1742 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1525 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1744 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1744 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1525 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1745 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#124 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#123 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#119 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1749 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1750 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1750 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#114 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#415 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1534 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1756 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1534 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#21 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#71 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#23 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#94 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#23 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1571 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#86 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1947 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#87 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1569 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#84 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#111 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#107 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#107 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#108 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#108 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#106 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#103 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#99 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#100 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#95 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#96 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#94 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#28 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#28 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#55 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#93 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#56 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#56 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#28 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#55 ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#56 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#417 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#290 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1540 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#421 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#420 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#289 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#289 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1541 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#418 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1541 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#288 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#418 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#288 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1761 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1542 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1542 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1763 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1543 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#422 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1542 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1545 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1547 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1547 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1547 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1770 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#30 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#27 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#25 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#28 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#25 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#26 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#52 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1771 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1768 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#40 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1773 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#58 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#34 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#35 FD: 41 BD: 5 +.+.: (wq_completion)bond0#30 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#57 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#41 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#37 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#37 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#33 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#34 ->&rq->__lock FD: 1 BD: 8 +.+.: ovs_mutex.wait_lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#34 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#51 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#144 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#427 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1796 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1557 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1557 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#76 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1556 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#426 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1797 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1778 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#292 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1781 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy32 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1777 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1781 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1780 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#292 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1779 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#15 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#15 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#15 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#30 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#30 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#29 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#29 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1777 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1777 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1773 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1775 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1781 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1789 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1784 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1789 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1775 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1785 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1775 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1771 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1906 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1774 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1774 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1770 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1776 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1778 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#29 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1784 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1780 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1784 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1785 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1782 FD: 158 BD: 26 +.+.: (wq_completion)hci0#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1554 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1792 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1554 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#57 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#60 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1554 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#14 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1806 ->&rq->__lock FD: 28 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1807 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1559 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1809 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1813 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1815 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1815 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1819 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#84 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1574 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1826 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#104 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#296 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#435 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1579 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#298 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#439 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1839 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#300 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1587 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#443 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1588 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1589 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1837 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#441 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#440 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#80 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#443 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#304 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#456 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#127 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#312 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#313 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#125 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#455 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#88 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#25 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1601 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_tx_wq#25 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#122 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#121 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#120 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#452 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1599 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1850 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1850 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#309 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#117 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1593 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#85 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1847 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#115 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#306 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#114 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1591 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#83 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#58 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#446 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#452 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#455 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#65 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1856 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1603 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1859 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1608 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#461 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1858 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1611 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1860 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1865 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1866 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1616 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#92 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1873 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1874 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#531 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1716 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1994 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1677 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1877 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1881 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1881 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#505 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1879 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#504 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#31 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1881 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#140 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#31 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#61 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1884 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1884 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#140 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1958 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1953 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#503 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1952 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#76 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#76 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1883 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#501 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1671 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1671 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1947 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#340 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#339 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1889 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1951 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1946 FD: 24 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#498 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#12 FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#12 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#23 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1889 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#320 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#93 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#467 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#466 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#466 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#466 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1891 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1892 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1893 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1892 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#73 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1897 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1622 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1890 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1891 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1899 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#103 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1668 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#104 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci4#11 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1625 FD: 133 BD: 46 +.+.: team->team_lock_key#34 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&c->lock ->&n->list_lock ->&rq->__lock ->lweventlist_lock ->remove_cache_srcu ->(console_sem).lock ->&____s->seqcount#2 ->&____s->seqcount FD: 41 BD: 1 +.+.: (wq_completion)bond0#34 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1940 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#21 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1666 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1946 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1941 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#102 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#495 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#33 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1668 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1667 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#11 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1667 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#11 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#103 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1951 FD: 1 BD: 49 +.+.: (wq_completion)phy61 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#72 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#102 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 49 +.+.: (wq_completion)phy62 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1942 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1947 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1899 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1905 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1903 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1943 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1650 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#20 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#334 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1665 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1666 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#31 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1711 FD: 24 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#5 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#144 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#143 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#144 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#142 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#143 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#8 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#32 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#491 FD: 184 BD: 1 +.+.: (wq_completion)hci0#11 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1663 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#489 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#332 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2003 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1661 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1660 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#488 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#486 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#30 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#30 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#100 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1935 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1655 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1654 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1565 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1928 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1644 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1921 FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg2#68 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#35 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#345 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#96 ->&rq->__lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#37 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#325 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#478 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1911 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#145 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#145 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1629 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1628 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1626 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#468 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#322 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#468 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1906 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1629 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1628 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1910 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1680 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1914 ->&rq->__lock FD: 1 BD: 2 +.+.: &net->ipv4.ra_mutex FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1912 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#323 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1639 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1639 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1918 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1913 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#322 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#472 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1637 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1636 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1636 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#18 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1635 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1916 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1909 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1913 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1910 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1636 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1912 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1914 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#322 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2001 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#79 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1640 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#1917 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1928 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1924 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1927 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1648 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#362 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#360 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#526 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1716 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1712 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1713 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1996 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#359 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#361 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#116 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#116 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#115 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#359 FD: 24 BD: 2 +.+.: (wq_completion)nfc12_nci_rx_wq#6 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#522 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1710 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1993 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1987 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1710 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1709 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1708 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1707 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1986 FD: 184 BD: 1 +.+.: (wq_completion)hci4#13 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1427 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1427 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1285 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1428 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1428 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1427 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#355 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1429 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1429 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1428 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1446 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1447 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1304 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1304 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1448 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1447 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1449 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1449 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1448 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1450 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1449 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1308 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1307 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1450 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1452 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1308 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1453 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1452 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1451 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1452 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1453 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1309 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1310 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1454 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1453 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1311 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1455 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1454 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1455 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1312 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1456 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1456 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1455 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1457 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1456 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1314 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1458 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1458 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1459 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1459 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1458 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1316 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1316 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc53_nci_tx_wq ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc53_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc53_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1481 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1659 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#47 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#18 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#390 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#47 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1659 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#60 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#57 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1480 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1479 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#63 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1479 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1658 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1657 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1478 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1478 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1653 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1630 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1630 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1629 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1653 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1652 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1651 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1632 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1649 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1649 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1633 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1478 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1477 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1648 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1634 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1633 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1634 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1646 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1635 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1635 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1634 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1645 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1635 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1636 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1644 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1638 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1638 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1644 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1639 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1638 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1640 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1640 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1639 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1643 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1641 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1641 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1642 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1641 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1643 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1642 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1647 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1647 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1647 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1648 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1648 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1650 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1649 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1650 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1652 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1651 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1652 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1654 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1653 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1655 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1656 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1655 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1656 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc54_nci_tx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc54_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc56_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc56_nci_cmd_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#2 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#3 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#3 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#8 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#8 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#8 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#37 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1660 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1660 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#48 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#48 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#264 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#264 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1485 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1484 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1673 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1675 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1677 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1676 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1677 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1816 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1678 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1677 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#54 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#26 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#26 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#52 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#51 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#51 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#26 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#53 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#52 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#51 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#52 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#53 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#54 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#53 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#2 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#7 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#36 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1679 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1678 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#263 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#47 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#47 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#47 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#263 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#263 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1484 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1483 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1680 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1679 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#2 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#48 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc54_nci_cmd_wq FD: 1 BD: 49 +.+.: (wq_completion)phy34 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1682 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1485 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1485 FD: 1 BD: 49 +.+.: (wq_completion)phy33 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1682 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1681 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1523 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1522 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1683 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1684 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1522 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#406 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#284 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1685 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1690 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1686 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1687 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1688 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1688 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1689 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1690 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1691 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1691 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#284 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1521 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1695 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#405 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1736 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#404 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1735 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1520 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1520 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#403 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#283 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#68 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#402 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1734 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1735 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1735 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1519 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1519 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1695 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1696 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1519 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#67 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#53 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#53 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1697 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1697 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#53 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#401 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#66 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#401 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#66 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#66 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1733 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1734 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#20 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1518 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#7 FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#52 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#52 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#281 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#52 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1518 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#281 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#399 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1517 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#51 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#51 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#51 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1516 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#65 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#65 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#65 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1732 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1733 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1733 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#280 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#280 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#279 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#398 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#279 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1515 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1515 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1515 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#278 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#11 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1487 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1486 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1486 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1698 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1698 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1513 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1513 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#277 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1513 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1732 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1699 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1699 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#1698 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#19 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#19 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#50 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#397 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#397 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#50 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1731 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#64 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1730 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1699 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1700 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1700 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1700 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#63 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1701 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1701 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#396 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#396 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1512 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#10 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1512 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1512 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1493 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1706 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1707 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1707 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1492 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1492 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1705 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1706 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1706 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1490 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1490 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1490 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#378 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1489 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#378 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1702 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1489 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1489 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#377 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#377 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1488 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1487 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1701 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1702 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1702 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#377 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1704 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1703 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1705 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#379 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1705 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1704 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1491 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1492 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1493 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#380 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1493 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1707 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1708 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#380 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#18 FD: 1 BD: 3 ....: &tags->lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#49 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#276 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#49 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#276 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#276 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#395 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1510 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1510 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#62 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#62 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1729 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1730 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#394 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#394 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1508 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1508 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1508 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#275 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#48 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#48 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#275 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#275 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#60 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#60 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1507 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1507 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1507 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#47 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#393 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#393 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#393 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#17 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1728 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1729 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#274 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#274 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#392 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#273 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#273 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#391 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#391 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#391 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1506 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1506 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1727 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1494 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1494 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1710 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1494 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1728 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#16 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#16 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#6 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#46 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#46 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#46 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1505 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1505 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1505 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#57 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#272 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#272 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#272 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#45 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#45 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#45 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1727 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#390 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#390 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#2 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#44 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#44 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#8 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#44 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#5 FD: 24 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#2 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#5 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1726 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#271 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1726 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#271 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#381 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#381 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1710 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1711 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1711 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1496 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1711 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1712 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1495 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1713 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1713 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1495 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1496 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1496 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1714 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1714 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1715 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1715 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1715 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1716 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1717 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1717 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1504 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#389 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#15 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#15 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#43 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#43 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#43 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#55 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1503 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#5 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#388 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#4 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#3 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1722 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#12 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1501 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1501 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1501 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#386 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#41 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#41 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#5 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#52 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#268 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#52 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#52 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1722 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1721 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#268 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#385 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1720 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1721 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#267 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#267 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#40 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#384 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#40 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#384 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#384 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#51 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1719 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1720 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1498 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1498 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#266 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#266 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1718 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#49 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#4 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#4 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#382 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#382 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#382 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#9 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1549 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1497 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1497 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1497 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1717 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#9 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1719 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1719 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#383 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#39 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#39 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1499 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#51 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#51 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#40 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1721 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#10 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#10 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1516 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1517 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1792 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1778 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1550 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1550 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1788 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1779 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1786 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#400 ->&rq->__lock ->&cfs_rq->removed.lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1734 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#12 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#67 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#67 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#402 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#404 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#69 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1739 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1523 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1524 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1524 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#410 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1743 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1743 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1526 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1526 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1527 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1745 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1746 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1746 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#411 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#411 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1528 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1746 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1747 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1747 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1748 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1748 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#125 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#125 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#124 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#122 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#122 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#121 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#121 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#120 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#120 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1748 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1749 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1530 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1530 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1530 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#119 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#118 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#117 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#118 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#117 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#116 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1749 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1531 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1531 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1531 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#115 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#113 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#113 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1583 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1583 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#415 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#415 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1755 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1756 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1534 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1752 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#71 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#21 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1533 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1533 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1533 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#412 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#412 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#412 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#286 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#286 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1750 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1751 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1532 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1532 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#54 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1532 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#13 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1752 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#21 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#71 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1753 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1755 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1755 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1754 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#72 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1754 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#72 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1535 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#414 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#414 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#414 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#287 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#413 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#413 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1753 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1536 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1536 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1584 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1756 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1757 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1538 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#439 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1538 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#439 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1582 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1835 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1582 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1583 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#299 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1834 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1833 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1837 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#57 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#57 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#78 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1580 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1580 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#78 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#297 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#297 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1579 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#436 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#436 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1830 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1834 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1834 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#435 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#435 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#77 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#77 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#77 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1829 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1833 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1833 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#296 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#296 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1828 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1832 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1832 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#434 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1576 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1827 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#295 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#295 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#433 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#433 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#433 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1826 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1575 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#112 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#111 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#432 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#432 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#110 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#108 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#432 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1829 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1574 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#107 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1574 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1824 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1575 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1828 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1827 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1573 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1573 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1822 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1826 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1758 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1758 ->&rq->__lock ->&cfs_rq->removed.lock FD: 158 BD: 26 +.+.: (wq_completion)hci0#10 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 184 BD: 1 +.+.: (wq_completion)hci0#9 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1539 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#416 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#416 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1539 ->&rq->__lock ->rcu_node_0 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1539 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1572 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1825 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1758 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1759 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1759 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1819 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1571 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#95 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1817 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#91 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#90 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#89 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1570 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy54 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1820 FD: 1 BD: 49 +.+.: (wq_completion)phy53 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#112 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#112 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#111 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#110 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#110 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#109 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#109 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#106 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#104 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#104 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#103 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#102 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#102 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#101 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#100 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#98 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#98 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#97 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#97 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#94 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#93 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#55 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#55 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1766 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1769 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1760 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1760 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#74 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#74 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#74 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#290 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#290 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#417 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#417 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1540 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1540 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1759 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1761 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1761 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#422 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1762 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1762 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#421 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#421 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1545 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1545 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#73 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#420 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#73 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#73 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#289 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1544 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1544 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#24 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1543 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1543 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1762 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1765 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#419 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#419 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1760 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1763 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1541 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#288 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1764 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1764 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#24 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1763 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1766 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1766 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1544 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1764 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1767 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1767 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1546 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1546 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1765 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1768 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1546 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1768 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#422 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#92 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#32 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1767 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1770 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#31 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#32 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#30 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#32 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#31 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#31 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#30 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#29 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#28 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#29 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#27 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#26 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#26 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#25 FD: 1 BD: 3 ....: rlock-AF_INET FD: 76 BD: 2 +.+.: sk_lock-AF_INET/1 ->slock-AF_INET#2 ->rlock-AF_INET ->&rq->__lock ->&list->lock#19 FD: 1 BD: 50 ....: key#24 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#91 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#91 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#90 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#90 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#89 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#57 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#56 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#55 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#54 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#49 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#47 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#46 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#45 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#44 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1771 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#43 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#42 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#102 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1548 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1773 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1769 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1772 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1772 ->&rq->__lock FD: 1 BD: 9 +.+.: nf_ct_proto_mutex.wait_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1548 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1548 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#39 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#38 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#58 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#57 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#56 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#56 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#55 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#55 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#54 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#49 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#49 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#48 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#47 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#47 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#46 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#44 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#43 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#42 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#42 FD: 158 BD: 26 +.+.: (wq_completion)hci4#10 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 184 BD: 1 +.+.: (wq_completion)hci4#9 ->&rq->__lock ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#41 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#40 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#39 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#38 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#38 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#36 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#36 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#35 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#35 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#33 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#39 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#33 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#36 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#37 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#53 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1557 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#427 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1799 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#76 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#76 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1793 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1797 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#292 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#426 FD: 1 BD: 49 +.+.: (wq_completion)phy52 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1785 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy51 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1786 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1786 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1783 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1787 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1787 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1550 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1791 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1791 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1792 FD: 184 BD: 1 +.+.: (wq_completion)hci0#7 ->&rq->__lock ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1551 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#425 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#425 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#30 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#29 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#29 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#58 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#58 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#57 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#30 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#60 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#60 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#59 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#59 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#30 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#59 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1553 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1791 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1795 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#424 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#424 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1552 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1790 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#56 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#14 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#56 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#56 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#83 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#82 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#81 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#67 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#66 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#65 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#64 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#83 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#82 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#81 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#81 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#80 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#79 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#79 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#78 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#78 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#77 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#77 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#76 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#75 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#75 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#73 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#72 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#71 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#70 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#70 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#69 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#69 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#68 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#67 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#67 FD: 1 BD: 83 ....: net_rwsem.wait_lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#66 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#66 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#64 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#63 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#62 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#60 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#60 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#59 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#74 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#59 FD: 1 BD: 52 +.+.: nf_hook_mutex.wait_lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#63 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#63 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#62 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#79 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#78 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#69 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#77 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#70 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#71 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#75 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#72 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#75 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#75 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#423 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#423 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1551 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1552 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1789 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1793 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#22 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#291 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#291 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1551 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#75 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#424 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1795 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#425 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1796 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1556 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1555 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1798 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1556 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1799 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1800 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1558 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1802 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1801 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1801 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1797 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#85 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#86 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#86 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1567 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1568 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1814 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1565 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1564 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1564 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1565 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1564 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#294 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#294 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#31 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#32 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#16 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#16 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#32 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#16 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1563 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1808 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#430 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1807 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1811 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1806 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1810 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1560 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1560 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1561 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#293 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1808 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#428 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1808 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1803 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1803 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1800 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1804 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1559 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1558 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1801 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1805 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1805 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1806 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1559 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1803 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1807 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1560 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#293 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1805 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1809 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#429 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1811 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#429 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1561 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1562 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#430 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1561 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#430 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#431 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1809 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1813 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1814 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1814 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1810 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1811 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1566 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1812 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1813 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1816 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1816 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1566 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1567 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1817 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1817 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1818 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1818 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#85 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1568 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1568 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1815 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1819 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1569 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#84 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#89 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#99 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#85 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1821 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1821 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#92 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#96 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#99 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#98 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1824 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1823 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#100 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1823 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1824 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1573 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#102 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1825 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#103 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1827 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#105 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#106 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1829 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1575 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1576 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1830 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1577 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1831 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1831 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1576 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1577 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1578 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1577 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1580 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1831 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1835 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#78 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1581 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1832 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1836 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1836 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#437 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#298 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1581 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1582 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#438 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#438 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#438 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1838 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#299 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1845 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#83 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#83 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#58 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#82 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#444 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1840 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1844 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1844 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#304 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#444 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#304 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#305 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1587 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1839 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1842 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1842 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#81 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#81 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1586 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1586 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1587 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#442 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#442 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1584 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1584 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1841 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1841 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#80 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#441 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#440 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#440 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1836 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1840 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1840 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#441 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#80 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1585 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1585 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#442 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1585 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1586 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#81 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#302 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#303 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#303 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#443 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1588 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1589 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1590 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1589 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#444 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#82 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#82 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1603 ->&rq->__lock FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#128 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1603 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#129 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1604 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1856 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#130 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#313 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#131 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#314 FD: 24 BD: 5 +.+.: (wq_completion)tipc_crypto#132 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#313 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#458 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#314 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#134 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#135 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1858 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#459 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1606 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#90 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1859 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#137 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1607 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1606 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1856 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1860 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1860 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1607 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1607 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1608 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#461 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1861 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1609 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#316 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#317 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1610 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1609 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#462 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#462 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1610 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#66 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#66 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#66 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#317 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1610 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#318 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1862 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1862 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#462 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1859 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1863 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1863 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1864 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1864 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1861 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1614 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1615 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1862 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1866 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1863 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1867 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1867 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1615 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1868 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1868 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1864 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#320 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1869 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1869 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci1#14 ->&rq->__lock ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 184 BD: 1 +.+.: (wq_completion)hci1#13 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1870 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1871 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1617 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1616 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#81 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1870 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1869 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1868 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1874 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1874 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#81 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1871 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#118 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1875 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1875 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#119 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#147 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1872 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1876 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1876 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#526 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#529 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1873 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1877 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1877 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#31 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1715 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1878 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1647 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1878 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1718 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1879 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1875 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1879 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#80 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1677 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1876 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1880 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1880 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1678 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1956 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1676 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1676 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#333 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1882 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1882 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1677 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#507 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1883 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1883 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#504 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#506 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#31 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#62 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#62 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#61 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 3 +.+.: &aux->poke_mutex FD: 32 BD: 2 +.+.: (work_completion)(&aux->work)#2 ->&aux->poke_mutex ->map_idr_lock ->&obj_hash[i].lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#61 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1880 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1885 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1885 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#139 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#139 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1954 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#139 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1963 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1675 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#140 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1675 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1676 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#503 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1960 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1674 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1951 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1675 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1949 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1955 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1950 ->&rq->__lock FD: 24 BD: 49 +.+.: (wq_completion)phy56 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#36 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1954 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1882 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1886 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1886 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#76 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#38 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy55 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1887 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1887 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1884 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1888 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1888 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#105 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#106 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#342 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#342 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#106 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#343 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1673 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1673 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#502 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1674 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#502 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1672 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1672 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1673 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#501 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#341 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#341 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#104 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#35 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#37 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#75 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#75 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#75 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#105 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#340 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#341 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1948 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1670 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1670 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1952 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#74 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#74 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1956 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#339 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1885 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1889 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1890 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#500 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#500 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#339 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#499 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1669 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#499 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#23 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#498 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#24 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1620 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#321 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#93 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#467 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#93 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1618 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1617 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1618 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1886 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1890 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1887 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1618 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1891 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#467 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#320 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1888 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1893 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1894 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#34 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#73 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#36 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1945 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1950 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1894 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1895 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1895 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1896 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1620 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1892 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1896 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1897 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1898 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1621 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1620 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1893 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1954 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1622 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1623 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1894 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1900 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1622 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#73 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#104 FD: 24 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#4 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1668 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#337 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1623 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1624 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1623 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1895 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1899 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1901 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1669 ->&rq->__lock FD: 158 BD: 26 +.+.: (wq_completion)hci4#12 ->&rq->__lock ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#337 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#338 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1625 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1625 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1626 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1896 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1902 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1900 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1901 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1897 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1903 ->&rq->__lock FD: 41 BD: 5 +.+.: (wq_completion)bond0#32 ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1915 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1904 FD: 1 BD: 1 +...: link_idr_lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1949 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1953 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#499 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#497 FD: 41 BD: 1 +.+.: (wq_completion)bond0#35 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1949 ->&rq->__lock FD: 133 BD: 46 +.+.: team->team_lock_key#35 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&rq->__lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->&c->lock ->&n->list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->(console_sem).lock ->lweventlist_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#100 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#10 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#32 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#63 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#64 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#63 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#64 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#493 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#493 ->&rq->__lock FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#32 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#64 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#493 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#71 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#71 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#71 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#334 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#33 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#20 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#33 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#4 FD: 24 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#3 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#333 FD: 855 BD: 13 +.+.: &devlink->lock_key#33 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&xa->xa_lock#19 ->&c->lock ->&n->list_lock ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rcu_node_0 ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->&____s->seqcount#2 ->&____s->seqcount ->stack_depot_init_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#4 FD: 24 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#4 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#3 FD: 24 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#2 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#2 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#2 FD: 24 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#2 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#2 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#4 FD: 184 BD: 1 +.+.: (wq_completion)hci2#15 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#143 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#142 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#142 FD: 855 BD: 13 +.+.: &devlink->lock_key#34 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&n->list_lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->&____s->seqcount#2 ->&____s->seqcount ->stack_depot_init_mutex ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#3 FD: 24 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#3 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#3 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#3 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#8 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#30 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#32 FD: 24 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#18 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1664 FD: 855 BD: 13 +.+.: &devlink->lock_key#35 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&n->list_lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->&____s->seqcount#2 ->&____s->seqcount ->stack_depot_init_mutex ->&rcu_state.expedited_wq ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 41 BD: 1 +.+.: (wq_completion)bond0#36 ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1663 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#70 FD: 133 BD: 46 +.+.: team->team_lock_key#36 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&rq->__lock ->&c->lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->lweventlist_lock ->&n->list_lock ->(console_sem).lock ->&____s->seqcount#2 ->&____s->seqcount FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#70 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#492 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#17 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#29 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#31 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#31 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#491 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#491 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1939 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1944 FD: 1 BD: 49 +.+.: (wq_completion)phy60 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1948 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#490 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#490 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1662 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#489 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#489 FD: 1 BD: 49 +.+.: (wq_completion)phy59 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#332 FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#33 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg2#66 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg2#65 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#33 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg1#66 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 49 +.+.: (wq_completion)phy38 FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg0#66 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg0#65 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg1#65 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#33 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 49 +.+.: (wq_completion)phy37 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1660 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1661 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#488 FD: 1 BD: 49 +.+.: (wq_completion)phy64 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1938 FD: 41 BD: 5 +.+.: (wq_completion)bond0#37 ->(work_completion)(&(&slave->notify_work)->work) FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1714 ->&rq->__lock FD: 855 BD: 13 +.+.: &devlink->lock_key#36 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_node_0 ->&(&fn_net->fib_chain)->lock ->stack_depot_init_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1943 FD: 24 BD: 49 +.+.: (wq_completion)phy63 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#536 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#331 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1946 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1945 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1660 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#487 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#1659 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#487 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1942 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1658 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1944 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#486 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#486 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1657 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1657 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1936 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1658 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1943 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#69 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#28 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#330 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#330 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#99 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#331 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#100 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1656 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1657 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1940 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#98 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1655 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#485 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#485 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1934 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1655 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1654 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1933 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1938 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1653 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1653 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1932 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1931 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#480 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#480 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1648 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1646 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1646 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1645 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1922 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1927 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1644 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1645 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1926 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1643 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1643 ->&rq->__lock FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#34 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#34 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg0#67 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 94 +.+.: rcu_state.exp_wake_mutex.wait_lock FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg1#67 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg1#68 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1925 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1644 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1920 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#327 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#327 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#328 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#95 FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg0#70 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg2#70 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#35 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2010 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg1#70 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg1#69 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg2#69 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#35 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#96 ->&rq->__lock FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#36 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg2#72 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg0#72 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#36 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg1#72 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg0#71 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg1#71 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1919 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1926 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1642 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1642 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1643 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#479 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#37 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#73 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#74 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#73 ->(work_completion)(&peer->transmit_handshake_work) ->rcu_node_0 ->&rq->__lock FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#74 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#73 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#479 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#479 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#94 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#94 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci3#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 26 +.+.: (wq_completion)hci3#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#325 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#326 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#478 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1641 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy57 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1906 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1631 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1913 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1632 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1630 FD: 1 BD: 49 +.+.: (wq_completion)phy65 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1905 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1631 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1910 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#146 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#146 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#145 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#146 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1903 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1908 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#469 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#469 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy66 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1627 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#469 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1628 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1627 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#321 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#321 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#468 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1626 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1901 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1627 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1902 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1909 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1907 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1630 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_tx_wq#470 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1631 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1632 FD: 1 BD: 49 +.+.: (wq_completion)phy58 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1632 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1907 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#325 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#477 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#324 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#476 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#476 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#476 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#475 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#475 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1638 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#323 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#472 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1635 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1633 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1634 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1908 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#472 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1911 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1916 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1917 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1919 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1921 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1638 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1920 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1922 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1915 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#477 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#477 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1640 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2002 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#147 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2003 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#527 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1641 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1916 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1921 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1923 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1922 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#478 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#326 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#326 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_rx_wq#95 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#327 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1646 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1645 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1924 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1929 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1931 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1649 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1925 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1932 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#480 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#97 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#98 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#68 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#98 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#68 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#483 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#483 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1652 ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#35 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#35 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#36 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#36 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#36 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#18 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#18 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#18 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#340 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#504 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#38 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1959 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1956 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1715 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1717 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1993 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1992 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1997 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2001 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#79 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#79 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#528 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#116 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1712 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1713 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1711 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1710 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1711 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1709 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1992 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#358 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#358 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#522 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#524 ->&rq->__lock FD: 158 BD: 1 +.+.: (wq_completion)hci4#14 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1708 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1985 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1990 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1705 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#357 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#357 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1706 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#359 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#520 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#521 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1992 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1704 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1982 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1987 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1991 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1703 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1703 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1704 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1702 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1702 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1703 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1980 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1985 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#356 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#518 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1701 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#519 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1702 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#521 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1700 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1700 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1701 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1978 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#115 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1699 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1699 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#355 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#355 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1977 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1982 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1986 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1981 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1697 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#354 FD: 1 BD: 1 ....: _rs.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#354 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1697 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#356 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#517 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#519 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#353 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1696 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1696 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#518 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1697 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#113 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#517 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#352 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#352 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#78 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#113 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#351 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#351 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1972 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1977 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1981 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1694 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1694 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1695 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1971 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1694 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1970 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1692 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1968 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1977 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#511 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1691 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#111 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#110 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#111 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1688 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1688 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1964 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1689 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1965 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1969 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#77 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#77 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#77 ->&rq->__lock FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#37 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#349 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1686 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#37 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#19 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#19 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#19 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#348 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1685 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#346 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#346 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#345 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1683 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#108 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#344 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#344 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#345 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#508 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#508 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#510 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#106 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1681 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#343 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1960 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1965 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1969 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#506 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1680 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1680 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1958 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1963 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg0#39 ->(work_completion)(&peer->transmit_handshake_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#39 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#40 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#40 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#39 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#20 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#20 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#40 ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#20 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#506 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1966 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1962 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1967 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1971 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1972 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#108 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1684 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1685 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1686 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1687 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#110 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci2#7 ->&rq->__lock ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 26 +.+.: (wq_completion)hci2#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#509 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#511 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1974 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1970 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1689 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#111 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1690 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1690 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1691 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#510 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1972 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1976 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#511 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1692 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1969 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1974 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1978 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1693 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1980 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1978 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1982 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#514 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#354 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1698 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#520 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#115 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1983 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1984 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1988 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1989 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1986 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1990 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1981 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1705 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1983 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1989 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1993 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1994 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1991 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1995 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#360 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1996 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#526 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1989 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1998 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1994 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#525 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1715 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_tx_wq#528 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#534 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#536 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2011 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2012 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2008 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1723 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2013 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1727 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2006 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2007 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy44 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2016 FD: 24 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#13 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#40 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2009 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#120 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#121 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#370 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#83 ->&rq->__lock FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg0#75 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#543 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2014 FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg1#41 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg0#42 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#42 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#21 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#21 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#41 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#21 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#42 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1221 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1362 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1362 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1221 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1222 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1363 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1222 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1222 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1223 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1364 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1223 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1223 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1224 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1365 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1365 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1224 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1225 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1225 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy40 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1225 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1226 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1226 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy39 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1227 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1227 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1228 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1229 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1370 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1229 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1229 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1230 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1230 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1231 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1232 FD: 164 BD: 1 +.+.: &uhid->devlock ->&rq->__lock ->&mm->mmap_lock ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&x->wait#9 ->&obj_hash[i].lock ->&c->lock ->semaphore->lock#3 ->&n->list_lock FD: 1 BD: 3 +.+.: dquirks_lock FD: 1 BD: 1 ....: &uhid->report_wait FD: 1 BD: 3 +.+.: &hdrv->dyn_lock FD: 1 BD: 4 ....: semaphore->lock#3 FD: 1 BD: 1 ....: &hdev->debug_wait FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1373 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1232 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1232 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1233 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#346 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#346 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#247 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#247 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1374 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1233 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1233 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1234 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1234 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1235 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1376 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1235 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1235 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1236 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#347 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#248 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#248 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#248 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1377 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1236 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1236 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1237 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1238 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#249 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1378 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#249 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1379 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1379 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#249 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1238 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1239 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1239 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1239 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1240 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1240 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1241 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1241 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1242 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1242 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1243 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1243 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1385 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1244 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1244 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1245 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1386 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1386 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1245 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1245 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1246 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1247 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1248 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#349 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#349 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#250 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#250 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1389 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1248 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1248 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1249 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1390 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#350 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1249 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#350 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#251 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1249 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#251 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1250 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#251 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1250 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1250 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1251 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1251 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1251 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1252 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1252 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1252 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1253 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#351 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#351 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1393 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1394 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1395 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1395 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1253 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1254 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1254 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1255 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#352 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#252 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#352 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#352 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#252 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#252 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1255 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1255 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1256 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1256 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1256 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1257 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1257 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1257 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1258 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1259 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1259 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1260 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1261 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1404 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1262 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1262 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1263 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1405 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1263 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1263 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1264 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#353 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#353 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1405 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1406 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1406 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#253 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#253 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#253 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1265 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1265 FD: 184 BD: 1 +.+.: (wq_completion)hci4#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 26 +.+.: (wq_completion)hci4#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1266 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1267 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#354 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#354 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1408 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1409 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1409 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1410 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1267 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1267 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1268 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1268 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1268 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1269 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1269 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1270 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1270 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1271 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1414 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1271 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1271 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1272 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1414 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1415 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1415 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1272 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1272 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1274 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1274 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1275 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1275 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1276 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1276 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1277 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1278 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1421 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1278 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1278 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1279 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1280 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1423 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1280 ->&rq->__lock FD: 1 BD: 49 +.+.: (wq_completion)phy31 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1281 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1281 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1282 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1282 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1282 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1283 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1283 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1283 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1284 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1284 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1284 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1285 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1285 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1286 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#355 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#355 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1286 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1286 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1287 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#254 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#254 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1430 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1287 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1288 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1289 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1289 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1290 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1433 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1290 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1290 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1291 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1291 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1291 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#356 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1292 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#356 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#356 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1435 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1435 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1292 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1292 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1293 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1437 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1437 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1293 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1293 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1294 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1438 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1438 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1294 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1294 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1295 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#357 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#357 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#357 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1295 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1295 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1296 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1438 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1439 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1439 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#255 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#255 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#255 FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg1#44 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 46 +.+.: (wq_completion)wg-kex-wg2#43 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#44 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#22 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#22 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#22 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1440 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1296 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1296 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1297 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1298 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1299 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1299 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1299 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1300 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1300 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1301 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1301 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1302 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1302 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1303 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1304 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1305 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1305 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1305 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1306 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1306 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1307 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1451 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1451 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#256 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#358 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#256 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#358 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#256 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1308 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1309 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1309 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1310 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1311 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1311 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1312 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1312 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1313 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1313 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1313 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1314 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1314 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1315 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1315 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1315 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1316 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1317 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1317 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1318 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1318 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1318 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1319 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1463 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1319 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1319 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1320 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1320 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1320 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1321 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1321 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1321 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1322 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1322 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1322 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1323 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1323 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1324 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1468 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1324 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1325 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1325 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1325 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1326 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1326 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1327 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1471 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1327 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1328 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1328 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1329 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1329 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1330 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1474 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1330 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1331 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1331 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1331 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#359 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1332 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#359 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1332 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1333 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1333 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1333 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1334 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1334 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1334 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1335 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1479 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1335 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1335 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1336 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1336 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1336 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1337 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1482 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1338 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1339 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1483 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1483 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1339 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1339 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1340 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1340 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1340 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1341 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1341 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1342 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#360 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#360 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#360 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#46 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#258 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#258 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#35 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#46 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#46 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1486 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1342 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1342 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1343 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1487 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1343 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1343 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1344 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1344 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1345 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1345 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1346 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1346 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1347 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1347 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1348 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1348 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1349 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1349 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1349 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1350 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1350 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1351 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1351 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1351 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1352 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1352 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1352 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1353 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1353 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1354 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#361 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1497 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1498 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1498 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1354 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1355 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1498 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1499 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1499 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1355 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1356 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1356 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1357 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1358 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1359 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1359 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1360 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1360 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1360 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1361 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1361 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1361 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1362 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1507 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#362 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#259 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#362 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#259 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#259 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#362 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1363 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1363 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1364 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1364 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1364 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1365 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1510 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1365 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1365 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1366 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1366 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1366 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1367 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1368 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1368 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1369 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1370 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1515 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1370 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1370 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1371 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#363 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#363 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1515 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1516 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1516 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1371 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1371 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1372 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#364 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1517 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1518 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1518 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1372 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1372 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1373 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1374 ->&rq->__lock FD: 51 BD: 1 .+.+: kn->active#53 ->&rq->__lock ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->khugepaged_wait.lock ->remove_cache_srcu ->&n->list_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1374 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1374 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1375 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1375 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1376 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1376 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1376 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1377 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#260 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1526 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1379 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1379 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1380 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1380 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1380 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1381 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#366 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#366 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#366 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1527 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1528 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1528 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1381 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1381 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1382 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1530 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1530 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1382 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1382 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1383 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1383 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1384 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#367 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#367 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1531 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1532 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1532 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1384 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1384 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1385 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1385 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1386 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#368 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#368 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#368 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#261 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#261 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#261 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1386 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1387 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1387 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1387 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1388 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1537 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1537 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1388 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1389 ->&rq->__lock FD: 184 BD: 1 +.+.: (wq_completion)hci1#9 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 158 BD: 26 +.+.: (wq_completion)hci1#10 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1389 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1390 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1390 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1540 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1391 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1392 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1541 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1392 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1392 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1393 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1542 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1542 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1393 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1393 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1394 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1543 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1394 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1394 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1395 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1544 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1544 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1395 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1395 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1396 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1396 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1396 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1397 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1546 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1397 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1397 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1398 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1398 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1398 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1399 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1399 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1400 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1401 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1402 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1463 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1403 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1403 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1404 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1404 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1404 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1405 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1406 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1595 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1443 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1444 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1444 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1445 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1446 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1598 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1446 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1446 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1447 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1447 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1448 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1448 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1448 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1449 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1449 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1449 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1450 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1450 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1451 FD: 158 BD: 26 +.+.: (wq_completion)hci4#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1603 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1451 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1452 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1452 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1452 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1453 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1453 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1454 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1454 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1455 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1607 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1607 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1455 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1456 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#372 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#372 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1607 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1608 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1608 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1609 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1456 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1457 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1457 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1458 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1611 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1458 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1458 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1459 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1612 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1459 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1460 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#373 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#373 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#262 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#262 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#262 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1613 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1460 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1461 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1461 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1461 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1462 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1463 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1464 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1464 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1465 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1465 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1466 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1466 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1466 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1467 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1467 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1467 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1468 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1469 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1469 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1469 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1470 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1471 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#374 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1623 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1624 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1624 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1625 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1471 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1471 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1472 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1473 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1473 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1474 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1475 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1475 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1475 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1476 FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg1#25 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg2#25 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 52 BD: 46 +.+.: (wq_completion)wg-kex-wg2#50 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 86 BD: 46 +.+.: (wq_completion)wg-crypt-wg0#25 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc49_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc49_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc48_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc48_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc48_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc47_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc47_nci_rx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc46_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc46_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc46_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc45_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc44_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc44_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc44_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc42_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc42_nci_rx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc42_nci_tx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc40_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc39_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc38_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_rx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc38_nci_tx_wq ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq FD: 24 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#407 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#408 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#285 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#285 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#285 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#70 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#409 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#70 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#409 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#409 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1524 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1740 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1740 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1739 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1741 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1740 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1742 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1741 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1742 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#410 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#410 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1525 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1526 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1745 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1527 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1527 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1744 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#411 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1528 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1528 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1747 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#126 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#126 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#127 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#127 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#128 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#128 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#129 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#129 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#130 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#130 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#131 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#131 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#132 FD: 24 BD: 5 +.+.: (wq_completion)tipc_send#132 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#133 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#133 FD: 24 BD: 5 +.+.: (wq_completion)tipc_rcv#134 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#134 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#135 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#135 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#136 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#136 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#137 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#137 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#138 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#138 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1845 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1841 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1591 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#445 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#445 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#445 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1590 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1590 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#23 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#23 FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#23 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#306 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#305 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#305 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1846 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#113 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1842 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1592 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#446 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1591 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#446 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#60 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#60 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#60 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#307 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#306 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#84 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#84 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#84 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#61 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#61 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#61 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#308 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#307 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1847 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#307 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1843 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#85 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#85 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#447 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#447 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#447 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#116 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1848 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1848 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1844 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#448 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#448 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#448 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1592 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1592 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#449 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#449 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#449 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1594 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1593 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1593 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#62 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#308 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#308 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#62 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#62 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1594 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1594 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#450 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#450 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#450 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1596 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1595 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#86 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#86 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#86 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1849 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1849 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1845 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1597 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1596 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1596 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1846 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1598 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1597 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#310 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1597 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#451 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#309 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#309 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#451 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#451 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1598 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1598 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#452 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#87 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#87 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#87 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1600 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1599 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1599 FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#311 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#310 FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#310 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1851 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1851 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1847 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#453 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#453 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#63 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#63 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1852 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#63 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1848 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1853 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1853 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1849 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#454 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#454 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#454 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1600 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1600 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_rx_wq#26 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#26 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_rx_wq#27 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#27 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#64 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#64 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#64 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#88 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#88 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#312 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#311 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#311 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1602 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1601 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1601 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#455 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1854 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1854 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1850 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1855 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1855 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1851 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#65 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#65 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#28 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#28 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#26 FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#8 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#8 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#8 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#456 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#456 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1602 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1602 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#15 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#15 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#89 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#89 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#89 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1852 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#457 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#457 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#457 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1857 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1857 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1853 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#458 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#314 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#458 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1605 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1604 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1604 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1858 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1854 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#459 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#459 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1605 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1605 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#316 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#136 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#90 FD: 24 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#90 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#315 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#315 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#138 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#460 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#460 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#460 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1606 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1608 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1861 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1857 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#316 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1609 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#91 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#91 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#91 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1612 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1611 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1611 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1613 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1612 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1612 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#319 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#463 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#463 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#463 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#318 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#318 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1614 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1613 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1613 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1614 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#464 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#464 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1615 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#465 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#29 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#29 ->rcu_node_0 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#465 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#27 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#465 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#319 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#319 FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#92 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#92 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#67 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#67 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#67 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1865 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1870 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1866 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1616 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1871 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1867 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#81 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2000 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1995 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1719 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1717 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1716 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#532 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#530 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#527 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#364 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#362 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#362 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#39 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#39 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#37 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#531 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#533 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#532 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#529 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1720 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1718 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1717 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#365 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2005 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#363 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#363 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1996 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1997 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1721 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#533 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#530 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1719 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1718 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#366 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#364 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1998 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1722 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1720 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1719 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#531 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2008 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2004 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1999 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2005 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2000 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2001 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1723 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#537 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#535 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#532 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1720 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2002 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1722 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1721 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2003 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#538 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1725 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#536 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#533 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1722 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#367 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#365 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#365 FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2009 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2004 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1724 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2014 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2010 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2005 FD: 24 BD: 49 +.+.: (wq_completion)phy75 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#540 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#538 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#535 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#25 FD: 24 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#24 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#13 FD: 24 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#82 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#82 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#82 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#40 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#38 ->&rq->__lock FD: 27 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#368 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#366 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#366 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#120 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#120 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#119 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#539 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1729 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1727 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1726 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2017 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2013 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2008 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1730 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1728 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1727 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#540 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#537 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2019 ->rcu_node_0 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2015 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#371 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#543 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#538 FD: 24 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#41 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#41 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#39 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2020 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#83 FD: 24 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#83 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2021 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1729 FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1728 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2011 ->&rq->__lock FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#544 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#542 FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#539 ->&rq->__lock FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg1#75 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg0#76 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2017 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2012 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1732 ->&rq->__lock FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1730 ->&rq->__lock FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg1#76 ->(work_completion)(&peer->transmit_handshake_work) FD: 85 BD: 1 +.+.: (wq_completion)wg-kex-wg2#75 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 86 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#38 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#1729 FD: 24 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#545 ->&rq->__lock FD: 52 BD: 1 +.+.: (wq_completion)wg-kex-wg2#76 ->(work_completion)(&peer->transmit_handshake_work) FD: 24 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#122 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#540 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#122 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#121 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#372 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#368 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#368 FD: 24 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#2023 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#2018 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#2013 FD: 24 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#1733 ->&rq->__lock FD: 24 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#1731 ->&rq->__lock all lock chains: irq_context: 0 (console_sem).lock irq_context: 0 &obj_hash[i].lock irq_context: 0 &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex irq_context: 0 fixmap_lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex irq_context: 0 console_mutex irq_context: 0 input_pool.lock irq_context: 0 base_crng.lock irq_context: 0 cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 crng_init_wait.lock irq_context: 0 early_pfn_lock irq_context: 0 devtree_lock irq_context: 0 resource_lock irq_context: 0 restart_handler_list.lock irq_context: 0 system_transition_mutex irq_context: 0 pcpu_lock irq_context: 0 debug_hook_lock irq_context: 0 zonelist_update_seq irq_context: 0 zonelist_update_seq zonelist_update_seq.seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &zone->lock irq_context: 0 &zone->lock &____s->seqcount irq_context: 0 &pcp->lock &zone->lock irq_context: 0 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &____s->seqcount irq_context: 0 pool_lock#2 irq_context: 0 pcpu_alloc_mutex irq_context: 0 pcpu_alloc_mutex pcpu_lock irq_context: 0 &n->list_lock irq_context: 0 &c->lock irq_context: 0 slab_mutex irq_context: 0 slab_mutex pool_lock#2 irq_context: 0 slab_mutex pcpu_alloc_mutex irq_context: 0 slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 trace_types_lock irq_context: 0 panic_notifier_list.lock irq_context: 0 die_chain.lock irq_context: 0 trace_event_sem irq_context: 0 batched_entropy_u32.lock irq_context: 0 batched_entropy_u32.lock crngs.lock irq_context: 0 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sysctl_lock irq_context: 0 &rq->__lock irq_context: 0 &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 init_task.pi_lock irq_context: 0 init_task.pi_lock &rq->__lock irq_context: 0 init_task.vtime_seqcount irq_context: 0 slab_mutex &c->lock irq_context: 0 slab_mutex &pcp->lock &zone->lock irq_context: 0 slab_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slab_mutex &____s->seqcount irq_context: 0 wq_pool_mutex irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 wq_pool_mutex &____s->seqcount irq_context: 0 wq_pool_mutex &c->lock irq_context: 0 wq_pool_mutex pool_lock#2 irq_context: 0 &wq->mutex irq_context: 0 &wq->mutex &pool->lock irq_context: 0 wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 shrinker_mutex irq_context: 0 rcu_node_0 irq_context: 0 rcu_state.barrier_lock irq_context: 0 rcu_state.barrier_lock rcu_node_0 irq_context: 0 &rnp->exp_poll_lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &pcp->lock &zone->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &____s->seqcount irq_context: 0 trace_event_sem trace_event_ida.xa_lock &c->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock pool_lock#2 irq_context: 0 trigger_cmd_mutex irq_context: 0 free_vmap_area_lock irq_context: 0 vmap_area_lock irq_context: 0 acpi_probe_mutex irq_context: 0 acpi_probe_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex free_vmap_area_lock irq_context: 0 acpi_probe_mutex vmap_area_lock irq_context: 0 acpi_probe_mutex &pcp->lock &zone->lock irq_context: 0 acpi_probe_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex &____s->seqcount irq_context: 0 acpi_probe_mutex init_mm.page_table_lock irq_context: 0 acpi_probe_mutex resource_lock irq_context: 0 acpi_probe_mutex &c->lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 acpi_probe_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 acpi_probe_mutex (console_sem).lock irq_context: 0 acpi_probe_mutex irq_domain_mutex irq_context: 0 acpi_probe_mutex pcpu_alloc_mutex irq_context: 0 acpi_probe_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_probe_mutex irq_domain_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 acpi_probe_mutex &domain->mutex irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_probe_mutex &domain->mutex pool_lock#2 irq_context: 0 acpi_probe_mutex &domain->mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex &desc->request_mutex irq_context: 0 acpi_probe_mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex cpu_pm_notifier.lock irq_context: 0 acpi_probe_mutex &obj_hash[i].lock irq_context: 0 acpi_probe_mutex purge_vmap_area_lock irq_context: 0 acpi_probe_mutex iort_msi_chip_lock irq_context: 0 acpi_probe_mutex &zone->lock irq_context: 0 acpi_probe_mutex &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex its_lock irq_context: 0 acpi_probe_mutex resource_lock irq_context: 0 acpi_probe_mutex efi_mem_reserve_persistent_lock irq_context: 0 acpi_probe_mutex lpi_range_lock irq_context: 0 acpi_probe_mutex syscore_ops_lock irq_context: 0 acpi_probe_mutex its_lock &its->lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex resource_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex resource_lock irq_context: 0 timekeeper_lock irq_context: 0 timekeeper_lock tk_core.seq.seqcount irq_context: 0 timekeeper_lock tk_core.seq.seqcount &obj_hash[i].lock irq_context: 0 acpi_probe_mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex (console_sem).lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock tk_core.seq.seqcount irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock tick_broadcast_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock jiffies_seq.seqcount irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 acpi_probe_mutex clocksource_mutex irq_context: 0 clockevents_lock irq_context: 0 tk_core.seq.seqcount irq_context: 0 &base->lock irq_context: 0 &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock irq_context: 0 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 batched_entropy_u64.lock irq_context: 0 batched_entropy_u64.lock crngs.lock irq_context: 0 pmus_lock irq_context: 0 pmus_lock pcpu_alloc_mutex irq_context: 0 pmus_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 pmus_lock pool_lock#2 irq_context: 0 pmus_lock &obj_hash[i].lock irq_context: 0 &swhash->hlist_mutex irq_context: 0 pmus_lock &cpuctx_mutex irq_context: 0 pmus_lock &obj_hash[i].lock pool_lock irq_context: 0 tty_ldiscs_lock irq_context: 0 console_lock irq_context: 0 console_lock pool_lock#2 irq_context: 0 console_lock &obj_hash[i].lock irq_context: 0 console_lock &pcp->lock &zone->lock irq_context: 0 console_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 console_lock &____s->seqcount irq_context: 0 console_lock &c->lock irq_context: 0 console_lock kbd_event_lock irq_context: 0 console_lock kbd_event_lock led_lock irq_context: 0 console_lock (console_sem).lock irq_context: 0 console_lock console_owner_lock irq_context: 0 init_task.alloc_lock irq_context: 0 acpi_ioremap_lock irq_context: 0 acpi_ioremap_lock pool_lock#2 irq_context: 0 semaphore->lock irq_context: 0 *(&acpi_gbl_reference_count_lock) irq_context: 0 hrtimer_bases.lock irq_context: 0 hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 percpu_counters_lock irq_context: 0 tomoyo_policy_lock irq_context: 0 tomoyo_policy_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex irq_context: 0 rtnl_mutex &c->lock irq_context: 0 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &____s->seqcount irq_context: 0 rtnl_mutex pool_lock#2 irq_context: 0 lock irq_context: 0 lock kernfs_idr_lock irq_context: 0 lock kernfs_idr_lock pool_lock#2 irq_context: 0 &root->kernfs_rwsem irq_context: 0 file_systems_lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 dq_list_lock irq_context: 0 sb_lock irq_context: 0 &type->s_umount_key/1 irq_context: 0 &type->s_umount_key/1 pool_lock#2 irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key/1 shrinker_mutex irq_context: 0 &type->s_umount_key/1 shrinker_mutex pool_lock#2 irq_context: 0 &type->s_umount_key/1 list_lrus_mutex irq_context: 0 &type->s_umount_key/1 sb_lock irq_context: 0 &type->s_umount_key/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key/1 &c->lock irq_context: 0 &type->s_umount_key/1 &____s->seqcount irq_context: 0 &type->s_umount_key/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key/1 percpu_counters_lock irq_context: 0 &type->s_umount_key/1 crngs.lock irq_context: 0 &type->s_umount_key/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key irq_context: 0 &type->s_umount_key/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &type->s_umount_key/1 &dentry->d_lock irq_context: 0 mnt_id_ida.xa_lock irq_context: 0 &dentry->d_lock irq_context: 0 mount_lock irq_context: 0 mount_lock mount_lock.seqcount irq_context: 0 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 irq_context: 0 &type->s_umount_key#2/1 pool_lock#2 irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#2/1 shrinker_mutex irq_context: 0 &type->s_umount_key#2/1 &c->lock irq_context: 0 &type->s_umount_key#2/1 &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#2/1 sb_lock irq_context: 0 &type->s_umount_key#2/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 irq_context: 0 &type->s_umount_key#2/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#2/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 &dentry->d_lock irq_context: 0 ucounts_lock irq_context: 0 proc_inum_ida.xa_lock irq_context: 0 init_fs.lock irq_context: 0 init_fs.lock init_fs.seq.seqcount irq_context: hardirq jiffies_lock irq_context: hardirq jiffies_lock jiffies_seq.seqcount irq_context: hardirq log_wait.lock irq_context: 0 &type->s_umount_key#3/1 irq_context: 0 &type->s_umount_key#3/1 pool_lock#2 irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#3/1 shrinker_mutex irq_context: 0 &type->s_umount_key#3/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#3/1 sb_lock irq_context: 0 &type->s_umount_key#3/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &c->lock irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#3/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#3/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 &dentry->d_lock irq_context: 0 &type->s_umount_key#3/1 &dentry->d_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 proc_subdir_lock irq_context: 0 proc_subdir_lock irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 &type->s_umount_key#4/1 irq_context: 0 &type->s_umount_key#4/1 pool_lock#2 irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#4/1 shrinker_mutex irq_context: 0 &type->s_umount_key#4/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#4/1 sb_lock irq_context: 0 &type->s_umount_key#4/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 irq_context: 0 &type->s_umount_key#4/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#4/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 &type->s_umount_key#4/1 &dentry->d_lock irq_context: 0 cgroup_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex &c->lock irq_context: 0 cgroup_mutex &____s->seqcount irq_context: 0 cgroup_mutex pool_lock#2 irq_context: 0 cgroup_mutex lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cgroup_mutex &obj_hash[i].lock irq_context: 0 cgroup_mutex cgroup_file_kn_lock irq_context: 0 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex css_set_lock irq_context: 0 lock cgroup_idr_lock irq_context: 0 lock cgroup_idr_lock pool_lock#2 irq_context: 0 cpuset_mutex irq_context: 0 cpuset_mutex callback_lock irq_context: 0 cgroup_mutex blkcg_pol_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex percpu_counters_lock irq_context: 0 cgroup_mutex shrinker_mutex irq_context: 0 cgroup_mutex shrinker_mutex pool_lock#2 irq_context: 0 cgroup_mutex &base->lock irq_context: 0 cgroup_mutex &base->lock &obj_hash[i].lock irq_context: 0 cgroup_mutex devcgroup_mutex irq_context: 0 cgroup_mutex cpu_hotplug_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 init_sighand.siglock irq_context: 0 init_mm.page_table_lock irq_context: 0 init_files.file_lock irq_context: 0 rcu_read_lock init_sighand.siglock irq_context: 0 lock pidmap_lock irq_context: 0 lock pidmap_lock pool_lock#2 irq_context: 0 pidmap_lock irq_context: 0 cgroup_threadgroup_rwsem irq_context: 0 cgroup_threadgroup_rwsem css_set_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock init_sighand.siglock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock irq_context: 0 &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &p->pi_lock irq_context: 0 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (kthreadd_done).wait.lock irq_context: 0 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock irq_context: 0 &p->alloc_lock irq_context: 0 &p->alloc_lock &____s->seqcount#2 irq_context: 0 fs_reclaim irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kthread_create_lock irq_context: 0 &x->wait irq_context: 0 rcu_read_lock &sighand->siglock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 &x->wait &p->pi_lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (worker)->lock irq_context: 0 wq_pool_mutex fs_reclaim irq_context: 0 wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 wq_pool_mutex kthread_create_lock irq_context: 0 wq_pool_mutex &p->pi_lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &x->wait irq_context: 0 wq_pool_mutex &rq->__lock irq_context: 0 wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex irq_context: 0 wq_mayday_lock irq_context: 0 &xa->xa_lock irq_context: 0 &pool->lock irq_context: 0 &pool->lock &p->pi_lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (&pool->mayday_timer) irq_context: 0 &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) &rnp->exp_poll_lock irq_context: 0 (null) irq_context: 0 (null) tk_core.seq.seqcount irq_context: 0 (&wq_watchdog_timer) irq_context: 0 (wq_completion)events_unbound irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) allocation_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq allocation_wait.lock irq_context: hardirq allocation_wait.lock &p->pi_lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 batched_entropy_u8.lock irq_context: 0 batched_entropy_u8.lock crngs.lock irq_context: 0 kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex irq_context: 0 rcu_tasks.cbs_gbl_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex irq_context: 0 rcu_tasks_trace.cbs_gbl_lock irq_context: hardirq rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#2 irq_context: 0 rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 rcu_tasks.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex kernel/rcu/tasks.h:152 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &ACCESS_PRIVATE(rtpcp, lock) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 (memory_chain).rwsem irq_context: 0 cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &rcu_state.gp_wq irq_context: 0 &stop_pi_lock irq_context: 0 &stop_pi_lock &rq->__lock irq_context: 0 &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &stopper->lock irq_context: 0 (module_notify_list).rwsem irq_context: 0 ddebug_lock irq_context: 0 iort_msi_chip_lock irq_context: 0 irq_domain_mutex irq_context: 0 irq_domain_mutex fs_reclaim irq_context: 0 irq_domain_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 irq_domain_mutex pool_lock#2 irq_context: 0 cci_probing irq_context: 0 cci_probing devtree_lock irq_context: 0 resource_lock irq_context: 0 fixmap_lock fs_reclaim irq_context: 0 fixmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 fixmap_lock &____s->seqcount irq_context: 0 fixmap_lock &c->lock irq_context: 0 fixmap_lock pool_lock#2 irq_context: 0 rcu_read_lock ptlock_ptr(ptdesc) irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex clockevents_lock irq_context: 0 watchdog_mutex irq_context: 0 watchdog_mutex cpu_hotplug_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 watchdog_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)events irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#5 irq_context: 0 watchdog_mutex cpu_hotplug_lock &x->wait#5 irq_context: 0 &newf->file_lock irq_context: 0 init_fs.lock &dentry->d_lock irq_context: 0 &p->vtime.seqcount irq_context: 0 cpu_hotplug_lock mem_hotplug_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.waiters.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpuset_hotplug_work irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &swhash->hlist_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock &cpuctx_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcp_batch_high_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &xa->xa_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &c->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &____s->seqcount irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock wq_pool_attach_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock (cpu_running).wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &base->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &base->lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&timer.timer) irq_context: softirq (&timer.timer) &p->pi_lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &rcu_state.gp_wq &p->pi_lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &obj_hash[i].lock irq_context: softirq rcu_callback pool_lock#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock cpu_hotplug_lock.waiters.lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: softirq rcu_callback cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 its_lock irq_context: 0 its_lock &its->lock irq_context: 0 clockevents_lock tick_broadcast_lock irq_context: 0 clockevents_lock jiffies_seq.seqcount irq_context: 0 clockevents_lock tk_core.seq.seqcount irq_context: 0 &irq_desc_lock_class irq_context: 0 &irq_desc_lock_class irq_controller_lock irq_context: 0 (cpu_running).wait.lock irq_context: 0 (cpu_running).wait.lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock (&timer.timer) irq_context: 0 &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock &irq_desc_lock_class irq_context: 0 cpu_hotplug_lock cpuhp_state-up &swhash->hlist_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock &cpuctx_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#4 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &obj_hash[i].lock pool_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &x->wait#7 irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &pool->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_node_0 irq_context: 0 cpu_hotplug_lock cpuhp_state-up resource_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up resource_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock irq_context: hardirq &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 &x->wait#6 &p->pi_lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 &x->wait#8 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work cpu_hotplug_lock irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 sched_domains_mutex irq_context: 0 sched_domains_mutex fs_reclaim irq_context: 0 sched_domains_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sched_domains_mutex pool_lock#2 irq_context: 0 sched_domains_mutex &obj_hash[i].lock irq_context: 0 sched_domains_mutex pcpu_alloc_mutex irq_context: 0 sched_domains_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sched_domains_mutex &c->lock irq_context: 0 sched_domains_mutex &____s->seqcount irq_context: 0 sched_domains_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sched_domains_mutex rcu_read_lock pool_lock#2 irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &cp->lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 sched_domains_mutex pcpu_lock irq_context: 0 slab_mutex fs_reclaim irq_context: 0 slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 irq_context: 0 &type->s_umount_key#5/1 fs_reclaim irq_context: 0 &type->s_umount_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 pool_lock#2 irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#5/1 shrinker_mutex irq_context: 0 &type->s_umount_key#5/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#5/1 sb_lock irq_context: 0 &type->s_umount_key#5/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#5/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#5/1 percpu_counters_lock irq_context: 0 &type->s_umount_key#5/1 crngs.lock irq_context: 0 &type->s_umount_key#5/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->s_umount_key#5/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#5/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#5/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->s_umount_key#5/1 &dentry->d_lock irq_context: 0 (setup_done).wait.lock irq_context: 0 namespace_sem irq_context: 0 namespace_sem fs_reclaim irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 namespace_sem &pcp->lock &zone->lock irq_context: 0 namespace_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 namespace_sem &____s->seqcount irq_context: 0 namespace_sem &c->lock irq_context: 0 namespace_sem pool_lock#2 irq_context: 0 namespace_sem mnt_id_ida.xa_lock irq_context: 0 namespace_sem pcpu_alloc_mutex irq_context: 0 namespace_sem pcpu_alloc_mutex pcpu_lock irq_context: 0 namespace_sem &dentry->d_lock irq_context: 0 namespace_sem mount_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &p->alloc_lock init_fs.lock irq_context: 0 rcu_read_lock &____s->seqcount#3 irq_context: 0 file_systems_lock irq_context: 0 &type->s_umount_key#6 irq_context: 0 &type->s_umount_key#6 fs_reclaim irq_context: 0 &type->s_umount_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#6 pool_lock#2 irq_context: 0 &type->s_umount_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &____s->seqcount irq_context: 0 &type->s_umount_key#6 &c->lock irq_context: 0 &type->s_umount_key#6 &lru->node[i].lock irq_context: 0 &type->s_umount_key#6 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#6 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key namespace_sem irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key namespace_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#2 irq_context: 0 rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_lock_key#5 irq_context: 0 &fs->lock irq_context: 0 &fs->lock &____s->seqcount#3 irq_context: 0 (setup_done).wait.lock &p->pi_lock irq_context: 0 req_lock irq_context: 0 of_mutex irq_context: 0 of_mutex fs_reclaim irq_context: 0 of_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 of_mutex pool_lock#2 irq_context: 0 of_mutex lock irq_context: 0 of_mutex lock kernfs_idr_lock irq_context: 0 of_mutex &root->kernfs_rwsem irq_context: 0 of_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &x->wait#9 irq_context: 0 &k->list_lock irq_context: 0 bus_type_sem irq_context: 0 &root->kernfs_rwsem irq_context: 0 &dev->power.lock irq_context: 0 dpm_list_mtx irq_context: 0 uevent_sock_mutex irq_context: 0 running_helpers_waitq.lock irq_context: 0 sysfs_symlink_target_lock irq_context: 0 &k->k_lock irq_context: 0 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex &dev->power.lock irq_context: 0 subsys mutex irq_context: 0 memory_blocks.xa_lock irq_context: 0 memory_blocks.xa_lock pool_lock#2 irq_context: 0 lock kernfs_idr_lock &c->lock irq_context: 0 lock kernfs_idr_lock &____s->seqcount irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (console_sem).lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#2 irq_context: 0 subsys mutex#3 irq_context: 0 dev_pm_qos_mtx irq_context: 0 dev_pm_qos_mtx fs_reclaim irq_context: 0 dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_mtx pool_lock#2 irq_context: 0 dev_pm_qos_mtx &dev->power.lock irq_context: 0 dev_pm_qos_mtx pm_qos_lock irq_context: 0 dev_pm_qos_sysfs_mtx irq_context: 0 dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 dev_pm_qos_sysfs_mtx lock irq_context: 0 dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 dev_pm_qos_mtx &c->lock irq_context: 0 dev_pm_qos_mtx &pcp->lock &zone->lock irq_context: 0 dev_pm_qos_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 dev_pm_qos_mtx &____s->seqcount irq_context: 0 register_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock fs_reclaim irq_context: 0 register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_lock pool_lock#2 irq_context: 0 register_lock proc_inum_ida.xa_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock &c->lock irq_context: 0 register_lock &____s->seqcount irq_context: 0 register_lock proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 &x->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_pm_notifier.lock irq_context: 0 (cpufreq_policy_notifier_list).rwsem irq_context: 0 (pm_chain_head).rwsem irq_context: 0 cpufreq_governor_mutex irq_context: 0 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 clocksource_mutex irq_context: 0 syscore_ops_lock irq_context: softirq &(&kfence_timer)->timer irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock irq_context: 0 binfmt_lock irq_context: 0 pin_fs_lock irq_context: 0 &type->s_umount_key#7/1 irq_context: 0 &type->s_umount_key#7/1 fs_reclaim irq_context: 0 &type->s_umount_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 pool_lock#2 irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#7/1 shrinker_mutex irq_context: 0 &type->s_umount_key#7/1 &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &c->lock irq_context: 0 &type->s_umount_key#7/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#7/1 sb_lock irq_context: 0 &type->s_umount_key#7/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#7/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 irq_context: 0 &type->s_umount_key#7/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#7/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#7/1 &dentry->d_lock irq_context: 0 rcu_read_lock mount_lock irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 irq_context: 0 &sb->s_type->i_mutex_key#2 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#8/1 irq_context: 0 &type->s_umount_key#8/1 fs_reclaim irq_context: 0 &type->s_umount_key#8/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 pool_lock#2 irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#8/1 shrinker_mutex irq_context: 0 &type->s_umount_key#8/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#8/1 sb_lock irq_context: 0 &type->s_umount_key#8/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#8/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 irq_context: 0 &type->s_umount_key#8/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#8/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &type->s_umount_key#8/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 chrdevs_lock irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 subsys mutex#4 irq_context: 0 async_lock irq_context: 0 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fwnode_link_lock irq_context: 0 regulator_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_maps_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_ww_class_mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex devtree_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &c->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (reboot_notifier_list).rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#5 &k->k_lock irq_context: 0 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pin_fs_lock irq_context: 0 purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_list_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex deferred_probe_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex rcu_read_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock irq_context: 0 &type->s_umount_key#9/1 irq_context: 0 &type->s_umount_key#9/1 fs_reclaim irq_context: 0 &type->s_umount_key#9/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 pool_lock#2 irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#9/1 shrinker_mutex irq_context: 0 &type->s_umount_key#9/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#9/1 sb_lock irq_context: 0 &type->s_umount_key#9/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &c->lock irq_context: 0 &type->s_umount_key#9/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 irq_context: 0 &type->s_umount_key#9/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#9/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &type->s_umount_key#9/1 &dentry->d_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 vmap_purge_lock irq_context: 0 vmap_purge_lock purge_vmap_area_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 &fp->aux->used_maps_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 proto_list_mutex irq_context: 0 targets_mutex irq_context: 0 nl_table_lock irq_context: 0 nl_table_wait.lock irq_context: 0 net_family_lock irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 rtnl_mutex fs_reclaim irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock irq_context: 0 sparse_irq_lock fs_reclaim irq_context: 0 sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock pool_lock#2 irq_context: 0 sparse_irq_lock &c->lock irq_context: 0 sparse_irq_lock &____s->seqcount irq_context: 0 sparse_irq_lock lock irq_context: 0 sparse_irq_lock lock kernfs_idr_lock irq_context: 0 sparse_irq_lock &root->kernfs_rwsem irq_context: 0 sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &cma->lock irq_context: 0 cma_mutex irq_context: 0 cma_mutex &zone->lock irq_context: 0 cma_mutex &zone->lock &____s->seqcount irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock &zone->lock irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cma_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cma_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cma_mutex &obj_hash[i].lock irq_context: 0 cma_mutex lock#2 irq_context: 0 &pool->lock#2 irq_context: 0 cma_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 freezer_lock irq_context: 0 audit_backlog_wait.lock irq_context: 0 &list->lock irq_context: 0 kauditd_wait.lock irq_context: 0 lock#3 irq_context: 0 lock#3 &zone->lock irq_context: 0 pcp_batch_high_lock irq_context: 0 khugepaged_mutex irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 gdp_mutex irq_context: 0 gdp_mutex &k->list_lock irq_context: 0 gdp_mutex fs_reclaim irq_context: 0 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 gdp_mutex pool_lock#2 irq_context: 0 gdp_mutex lock irq_context: 0 gdp_mutex lock kernfs_idr_lock irq_context: 0 gdp_mutex &root->kernfs_rwsem irq_context: 0 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 subsys mutex#6 irq_context: 0 subsys mutex#6 &k->k_lock irq_context: 0 subsys mutex#7 irq_context: 0 subsys mutex#7 &k->list_lock irq_context: 0 subsys mutex#7 &k->k_lock irq_context: 0 regmap_debugfs_early_lock irq_context: 0 (acpi_reconfig_chain).rwsem irq_context: 0 __i2c_board_lock irq_context: 0 quarantine_lock irq_context: 0 core_lock irq_context: 0 core_lock &k->list_lock irq_context: 0 core_lock &k->k_lock irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu quarantine_lock irq_context: 0 nl_table_lock irq_context: 0 thermal_governor_lock irq_context: 0 thermal_governor_lock thermal_list_lock irq_context: 0 cpuidle_lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpuidle_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 cpuidle_lock &obj_hash[i].lock irq_context: 0 cpuidle_lock (console_sem).lock irq_context: 0 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_lock_key#8 irq_context: 0 &dir->lock irq_context: 0 k-sk_lock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR k-slock-AF_QIPCRTR irq_context: 0 k-slock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_ports.xa_lock irq_context: 0 k-sk_lock-AF_QIPCRTR pool_lock#2 irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_node_lock irq_context: 0 k-sk_lock-AF_QIPCRTR &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 resource_lock pool_lock#2 irq_context: 0 resource_lock &obj_hash[i].lock irq_context: 0 resource_lock &c->lock irq_context: 0 resource_lock &____s->seqcount irq_context: 0 crngs.lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 tty_mutex irq_context: 0 iova_cache_mutex irq_context: 0 iova_cache_mutex cpu_hotplug_lock irq_context: 0 iova_cache_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 iova_cache_mutex slab_mutex irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 iova_cache_mutex slab_mutex pool_lock#2 irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 gdp_mutex &c->lock irq_context: 0 gdp_mutex &____s->seqcount irq_context: 0 subsys mutex#8 irq_context: 0 subsys mutex#8 &k->k_lock irq_context: 0 device_links_lock irq_context: 0 uidhash_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock irq_context: 0 rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 oom_reaper_wait.lock irq_context: 0 subsys mutex#9 irq_context: 0 &pgdat->kcompactd_wait irq_context: 0 hugetlb_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) pool_lock#2 irq_context: 0 memory_tier_lock irq_context: 0 memory_tier_lock fs_reclaim irq_context: 0 memory_tier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 memory_tier_lock pool_lock#2 irq_context: 0 memory_tier_lock &x->wait#9 irq_context: 0 memory_tier_lock &obj_hash[i].lock irq_context: 0 memory_tier_lock &k->list_lock irq_context: 0 memory_tier_lock &c->lock irq_context: 0 memory_tier_lock &____s->seqcount irq_context: 0 memory_tier_lock lock irq_context: 0 memory_tier_lock lock kernfs_idr_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 memory_tier_lock bus_type_sem irq_context: 0 memory_tier_lock sysfs_symlink_target_lock irq_context: 0 memory_tier_lock &k->k_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &dev->power.lock irq_context: 0 memory_tier_lock dpm_list_mtx irq_context: 0 memory_tier_lock uevent_sock_mutex irq_context: 0 memory_tier_lock running_helpers_waitq.lock irq_context: 0 memory_tier_lock &dev->mutex &k->list_lock irq_context: 0 memory_tier_lock &dev->mutex &k->k_lock irq_context: 0 memory_tier_lock &dev->mutex &dev->power.lock irq_context: 0 memory_tier_lock subsys mutex#10 irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 memory_tier_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &rcu_state.expedited_wq &p->pi_lock irq_context: 0 khugepaged_mutex fs_reclaim irq_context: 0 khugepaged_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 khugepaged_mutex pool_lock#2 irq_context: 0 khugepaged_mutex kthread_create_lock irq_context: 0 khugepaged_mutex &p->pi_lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 khugepaged_mutex &x->wait irq_context: 0 khugepaged_mutex &rq->__lock irq_context: 0 khugepaged_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ksm_thread_mutex irq_context: 0 ksm_thread_wait.lock irq_context: 0 khugepaged_mutex &obj_hash[i].lock irq_context: 0 lock#2 irq_context: 0 khugepaged_mm_lock irq_context: 0 khugepaged_wait.lock irq_context: 0 khugepaged_mutex lock#3 irq_context: 0 khugepaged_mutex lock#3 &zone->lock irq_context: 0 khugepaged_mutex pcp_batch_high_lock irq_context: 0 cgroup_mutex fs_reclaim irq_context: 0 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_ops_lock irq_context: 0 crypto_alg_sem irq_context: 0 crypto_alg_sem (crypto_chain).rwsem irq_context: 0 cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_hotplug_lock &wq->mutex irq_context: 0 cpu_hotplug_lock &wq->mutex &pool->lock irq_context: 0 cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &x->wait irq_context: 0 cpu_hotplug_lock &rq->__lock irq_context: 0 cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &____s->seqcount irq_context: 0 bio_slab_lock irq_context: 0 bio_slab_lock fs_reclaim irq_context: 0 bio_slab_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex irq_context: 0 bio_slab_lock slab_mutex fs_reclaim irq_context: 0 bio_slab_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock slab_mutex pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex &c->lock irq_context: 0 bio_slab_lock slab_mutex &____s->seqcount irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock pool_lock#2 irq_context: 0 major_names_lock irq_context: 0 major_names_lock fs_reclaim irq_context: 0 major_names_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 major_names_lock pool_lock#2 irq_context: 0 major_names_lock major_names_spinlock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &rsp->gp_wait irq_context: 0 rcu_tasks.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&pool->mayday_timer) &pool->lock irq_context: softirq (&pool->mayday_timer) &pool->lock wq_mayday_lock irq_context: softirq (&pool->mayday_timer) &obj_hash[i].lock irq_context: softirq (&pool->mayday_timer) &base->lock irq_context: softirq (&pool->mayday_timer) &base->lock &obj_hash[i].lock irq_context: softirq (&rtpcp->lazy_timer) irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &base->lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) (&timer.timer) irq_context: 0 rcu_tasks.tasks_gp_mutex (console_sem).lock irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_lock irq_context: softirq drivers/char/random.c:251 irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (next_reseed).work irq_context: 0 (wq_completion)events_unbound (next_reseed).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work input_pool.lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work base_crng.lock irq_context: 0 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: softirq mm/vmstat.c:2022 irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock &obj_hash[i].lock irq_context: hardirq allocation_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 slab_mutex &rq->__lock irq_context: 0 &pool->lock &base->lock irq_context: 0 &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 slab_mutex batched_entropy_u8.lock irq_context: 0 slab_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 slab_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 slab_mutex kfence_freelist_lock irq_context: 0 console_lock fs_reclaim irq_context: 0 console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock &x->wait#9 irq_context: 0 console_lock &k->list_lock irq_context: 0 console_lock gdp_mutex irq_context: 0 console_lock gdp_mutex &k->list_lock irq_context: 0 console_lock gdp_mutex fs_reclaim irq_context: 0 console_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock gdp_mutex pool_lock#2 irq_context: 0 console_lock gdp_mutex lock irq_context: 0 console_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock lock irq_context: 0 console_lock lock kernfs_idr_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock bus_type_sem irq_context: 0 console_lock sysfs_symlink_target_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &dev->power.lock irq_context: 0 console_lock dpm_list_mtx irq_context: 0 console_lock uevent_sock_mutex irq_context: 0 console_lock running_helpers_waitq.lock irq_context: 0 console_lock subsys mutex#11 irq_context: 0 console_lock subsys mutex#11 &k->k_lock irq_context: 0 shrink_qlist.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 &ACCESS_PRIVATE(sdp, lock) irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &x->wait#3 irq_context: softirq &(&ssp->srcu_sup->work)->timer irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock irq_context: 0 cpu_hotplug_lock flush_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock (work_completion)(&sfw->work) irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock (wq_completion)slub_flushwq irq_context: 0 cpu_hotplug_lock flush_lock &x->wait#10 irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)slub_flushwq irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &c->lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &n->list_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &obj_hash[i].lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock pool_lock irq_context: 0 &meta->lock irq_context: 0 &n->list_lock &c->lock irq_context: 0 clk_debug_lock irq_context: 0 clocks_mutex irq_context: 0 acpi_scan_lock irq_context: 0 acpi_scan_lock semaphore->lock irq_context: 0 acpi_scan_lock fs_reclaim irq_context: 0 acpi_scan_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock pool_lock#2 irq_context: 0 acpi_scan_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &x->wait#9 irq_context: 0 acpi_scan_lock acpi_device_lock irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_device_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock &xa->xa_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_device_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &k->list_lock irq_context: 0 acpi_scan_lock lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock bus_type_sem irq_context: 0 acpi_scan_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &k->k_lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &c->lock irq_context: 0 acpi_scan_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->power.lock irq_context: 0 acpi_scan_lock dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock subsys mutex#12 irq_context: 0 acpi_scan_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_device_lock &c->lock irq_context: 0 acpi_scan_lock acpi_device_lock &____s->seqcount irq_context: 0 acpi_scan_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &rq->__lock irq_context: 0 acpi_scan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &n->list_lock irq_context: 0 acpi_scan_lock &n->list_lock &c->lock irq_context: 0 acpi_scan_lock quarantine_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_ioremap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_ioremap_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_ioremap_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_ioremap_lock vmap_area_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &device->physical_node_lock pool_lock#2 irq_context: 0 acpi_scan_lock &device->physical_node_lock lock irq_context: 0 acpi_scan_lock &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &device->physical_node_lock &c->lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &____s->seqcount irq_context: 0 acpi_scan_lock irq_domain_mutex irq_context: 0 acpi_scan_lock &domain->mutex irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &c->lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &domain->mutex &irq_desc_lock_class irq_context: 0 acpi_scan_lock &domain->mutex fs_reclaim irq_context: 0 acpi_scan_lock &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &domain->mutex pool_lock#2 irq_context: 0 acpi_scan_lock resource_lock irq_context: 0 acpi_scan_lock &(&priv->bus_notifier)->rwsem irq_context: 0 acpi_scan_lock &(&priv->bus_notifier)->rwsem &device->physical_node_lock irq_context: 0 acpi_scan_lock fwnode_link_lock irq_context: 0 acpi_scan_lock fwnode_link_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &device->physical_node_lock irq_context: 0 acpi_scan_lock &dev->mutex device_links_srcu irq_context: 0 acpi_scan_lock &dev->mutex fwnode_link_lock irq_context: 0 acpi_scan_lock &dev->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex &dev->devres_lock irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_maps_mutex irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 acpi_scan_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_wakeup_lock irq_context: 0 acpi_scan_lock &dev->mutex resource_lock irq_context: 0 acpi_scan_lock &dev->mutex free_vmap_area_lock irq_context: 0 acpi_scan_lock &dev->mutex vmap_area_lock irq_context: 0 acpi_scan_lock &dev->mutex init_mm.page_table_lock irq_context: 0 acpi_scan_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &dev->mutex chrdevs_lock irq_context: 0 acpi_scan_lock &dev->mutex tty_mutex irq_context: 0 acpi_scan_lock &dev->mutex proc_subdir_lock irq_context: 0 acpi_scan_lock &dev->mutex proc_inum_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex proc_subdir_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex device_links_srcu irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fwnode_link_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->devres_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex subsys mutex#14 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &xa->xa_lock#3 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)pm irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex (console_sem).lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex syslog_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex (console_sem).lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) pcpu_lock irq_context: softirq rcu_callback &x->wait#24 irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock irq_context: softirq &(&group->avgs_work)->timer irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/memcontrol.c:679 irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock crngs.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex kfence_freelist_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex ctrl_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex semaphore->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#15 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &n->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &n->list_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->k_lock klist_remove_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex req_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &x->wait#11 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers irq_context: 0 sb_writers mount_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &x->wait#11 irq_context: 0 &x->wait#11 &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#16 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#16 &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex chrdevs_lock irq_context: 0 acpi_scan_lock &dev->mutex fwnode_link_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex probe_waitqueue.lock irq_context: 0 acpi_scan_lock subsys mutex#4 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock (console_sem).lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock free_vmap_area_lock irq_context: 0 acpi_scan_lock vmap_area_lock irq_context: 0 acpi_scan_lock init_mm.page_table_lock irq_context: 0 acpi_scan_lock io_range_mutex irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock gdp_mutex irq_context: 0 acpi_scan_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock gdp_mutex lock irq_context: 0 acpi_scan_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock subsys mutex#17 irq_context: 0 acpi_scan_lock subsys mutex#17 &k->k_lock irq_context: 0 acpi_scan_lock acpi_hp_context_lock irq_context: 0 acpi_scan_lock acpi_hp_context_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_hp_context_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_hp_context_lock pool_lock#2 irq_context: 0 acpi_scan_lock bridge_mutex irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock pci_lock irq_context: 0 acpi_scan_lock pci_acpi_companion_lookup_sem irq_context: 0 acpi_scan_lock pci_slot_mutex irq_context: 0 acpi_scan_lock resource_alignment_lock irq_context: 0 acpi_scan_lock device_links_srcu irq_context: 0 acpi_scan_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 acpi_scan_lock iort_msi_chip_lock irq_context: 0 acpi_scan_lock subsys mutex#18 irq_context: 0 acpi_scan_lock devtree_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock batched_entropy_u8.lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock kfence_freelist_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock acpi_link_lock irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_link_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_link_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_link_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_link_lock (console_sem).lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock &c->lock irq_context: 0 acpi_scan_lock acpi_link_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_dep_list_lock irq_context: 0 acpi_scan_lock power_resource_list_lock irq_context: 0 acpi_device_lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: 0 klist_remove_lock irq_context: 0 &k->k_lock klist_remove_lock irq_context: 0 kernfs_idr_lock irq_context: 0 console_lock console_srcu console_owner_lock irq_context: 0 console_lock console_srcu console_owner irq_context: 0 console_lock console_srcu console_owner &port_lock_key irq_context: 0 console_lock console_srcu console_owner console_owner_lock irq_context: 0 k-sk_lock-AF_NETLINK irq_context: 0 k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: softirq (&timer.timer) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 k-slock-AF_NETLINK irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#10/1 irq_context: 0 &type->s_umount_key#10/1 fs_reclaim irq_context: 0 &type->s_umount_key#10/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 pool_lock#2 irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#10/1 shrinker_mutex irq_context: 0 &type->s_umount_key#10/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#10/1 sb_lock irq_context: 0 &type->s_umount_key#10/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#10/1 &____s->seqcount irq_context: 0 &type->s_umount_key#10/1 &c->lock irq_context: 0 &type->s_umount_key#10/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 irq_context: 0 &type->s_umount_key#10/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#10/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#10/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 irq_context: 0 &type->s_umount_key#11/1 fs_reclaim irq_context: 0 &type->s_umount_key#11/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 pool_lock#2 irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#11/1 shrinker_mutex irq_context: 0 &type->s_umount_key#11/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#11/1 sb_lock irq_context: 0 &type->s_umount_key#11/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#11/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 irq_context: 0 &type->s_umount_key#11/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#11/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#11/1 &c->lock irq_context: 0 &type->s_umount_key#11/1 &____s->seqcount irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 &dentry->d_lock irq_context: 0 &mm->mmap_lock irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start dma_fence_map irq_context: 0 &mm->mmap_lock irq_context: 0 key irq_context: 0 attribute_container_mutex irq_context: 0 triggers_list_lock irq_context: 0 leds_list_lock irq_context: 0 bus_type_sem irq_context: 0 (usb_notifier_list).rwsem irq_context: 0 &device->physical_node_lock irq_context: 0 rc_map_lock irq_context: 0 &root->kernfs_rwsem &____s->seqcount irq_context: 0 pci_lock irq_context: 0 subsys mutex#19 irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 (efi_runtime_lock).lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) cpu_asid_lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) efi_rt_lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock irq_context: 0 (efivars_lock).lock irq_context: 0 devfreq_list_lock irq_context: 0 &entry->access irq_context: 0 info_mutex irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 info_mutex fs_reclaim irq_context: 0 info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 info_mutex &c->lock irq_context: 0 info_mutex &____s->seqcount irq_context: 0 info_mutex pool_lock#2 irq_context: 0 info_mutex proc_inum_ida.xa_lock irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex running_helpers_waitq.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock nl_table_lock irq_context: 0 rcu_read_lock nl_table_wait.lock irq_context: 0 qdisc_mod_lock irq_context: 0 bt_proto_lock irq_context: 0 hci_cb_list_lock irq_context: 0 mgmt_chan_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rate_ctrl_mutex irq_context: 0 rate_ctrl_mutex fs_reclaim irq_context: 0 rate_ctrl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rate_ctrl_mutex pool_lock#2 irq_context: 0 netlbl_domhsh_lock irq_context: 0 netlbl_unlhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock pool_lock#2 irq_context: 0 misc_mtx irq_context: 0 misc_mtx fs_reclaim irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx pool_lock#2 irq_context: 0 misc_mtx &x->wait#9 irq_context: 0 misc_mtx &obj_hash[i].lock irq_context: 0 misc_mtx &____s->seqcount irq_context: 0 misc_mtx &k->list_lock irq_context: 0 misc_mtx gdp_mutex irq_context: 0 misc_mtx gdp_mutex &k->list_lock irq_context: 0 misc_mtx gdp_mutex fs_reclaim irq_context: 0 misc_mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx gdp_mutex pool_lock#2 irq_context: 0 misc_mtx gdp_mutex lock irq_context: 0 misc_mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx lock irq_context: 0 misc_mtx lock kernfs_idr_lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx bus_type_sem irq_context: 0 misc_mtx sysfs_symlink_target_lock irq_context: 0 misc_mtx &c->lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->power.lock irq_context: 0 misc_mtx dpm_list_mtx irq_context: 0 misc_mtx req_lock irq_context: 0 misc_mtx &p->pi_lock irq_context: 0 misc_mtx &x->wait#11 irq_context: 0 misc_mtx &rq->__lock irq_context: 0 misc_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx uevent_sock_mutex irq_context: 0 misc_mtx running_helpers_waitq.lock irq_context: 0 misc_mtx subsys mutex#21 irq_context: 0 misc_mtx subsys mutex#21 &k->k_lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex irq_context: 0 input_mutex irq_context: 0 input_mutex input_devices_poll_wait.lock irq_context: 0 (netlink_chain).rwsem irq_context: 0 proto_tab_lock irq_context: 0 random_ready_notifier.lock irq_context: 0 random_ready_notifier.lock crngs.lock irq_context: 0 misc_mtx misc_minors_ida.xa_lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &rq->__lock &rt_rq->rt_runtime_lock irq_context: 0 wtd_deferred_reg_mutex irq_context: 0 &type->s_umount_key#12/1 irq_context: 0 &type->s_umount_key#12/1 fs_reclaim irq_context: 0 &type->s_umount_key#12/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 pool_lock#2 irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#12/1 shrinker_mutex irq_context: 0 &type->s_umount_key#12/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#12/1 sb_lock irq_context: 0 &type->s_umount_key#12/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#12/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 irq_context: 0 &type->s_umount_key#12/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#12/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#12/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#11 irq_context: 0 clocksource_mutex cpu_hotplug_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 clocksource_mutex (console_sem).lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner_lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner irq_context: 0 clocksource_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 clocksource_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#13/1 irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#13/1 fs_reclaim irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#13/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#13/1 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &c->lock irq_context: 0 &type->s_umount_key#13/1 shrinker_mutex irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &____s->seqcount irq_context: 0 &type->s_umount_key#13/1 list_lrus_mutex irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#13/1 sb_lock irq_context: 0 &type->s_umount_key#13/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#13/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#13/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#13/1 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &type->s_umount_key#13/1 &____s->seqcount irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &type->s_umount_key#13/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#13/1 &dentry->d_lock irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->page_table_lock irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &sb->s_type->i_mutex_key#5 irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sb->s_type->i_mutex_key#5 rename_lock.seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#5 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex fs_reclaim irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#5 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#5 &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex init_fs.lock irq_context: 0 &sb->s_type->i_mutex_key#5 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 irq_context: 0 &type->s_umount_key#15/1 irq_context: 0 &type->s_umount_key#15/1 fs_reclaim irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#15/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &type->s_umount_key#15/1 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 &type->s_umount_key#15/1 shrinker_mutex irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->s_umount_key#15/1 list_lrus_mutex irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->s_umount_key#15/1 sb_lock irq_context: 0 &type->s_umount_key#15/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#15/1 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#15/1 &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#15/1 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sig->cred_guard_mutex &folio_wait_table[i] irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 &dentry->d_lock irq_context: 0 &type->s_umount_key#15/1 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_wait.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock irq_context: 0 &type->s_umount_key#16/1 irq_context: 0 &type->s_umount_key#16/1 fs_reclaim irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#16/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#16/1 pool_lock#2 irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#16/1 shrinker_mutex irq_context: 0 &type->s_umount_key#16/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#16/1 sb_lock irq_context: 0 &type->s_umount_key#16/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#16/1 &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &type->s_umount_key#16/1 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &type->s_umount_key#16/1 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 irq_context: 0 &sig->cred_guard_mutex binfmt_lock irq_context: 0 &type->s_umount_key#16/1 &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex init_binfmt_misc.entries_lock irq_context: 0 &type->s_umount_key#16/1 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#16/1 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 irq_context: 0 &type->s_umount_key#17/1 fs_reclaim irq_context: 0 &type->s_umount_key#17/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 &c->lock irq_context: 0 &type->s_umount_key#17/1 &____s->seqcount irq_context: 0 &type->s_umount_key#17/1 pool_lock#2 irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#17/1 shrinker_mutex irq_context: 0 &type->s_umount_key#17/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#17/1 sb_lock irq_context: 0 &type->s_umount_key#17/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#17/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 irq_context: 0 &type->s_umount_key#17/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 &type->s_umount_key#17/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#15 irq_context: hardirq tick_broadcast_lock irq_context: hardirq tick_broadcast_lock jiffies_lock irq_context: hardirq|softirq &rt_b->rt_runtime_lock irq_context: hardirq|softirq &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: hardirq|softirq &rt_rq->rt_runtime_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events timer_update_work irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 kclist_lock irq_context: 0 kclist_lock resource_lock irq_context: 0 kclist_lock fs_reclaim irq_context: 0 kclist_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kclist_lock pool_lock#2 irq_context: 0 &type->s_umount_key#18/1 irq_context: 0 &type->s_umount_key#18/1 fs_reclaim irq_context: 0 &type->s_umount_key#18/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 pool_lock#2 irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#18/1 shrinker_mutex irq_context: 0 &type->s_umount_key#18/1 &c->lock irq_context: 0 &type->s_umount_key#18/1 &____s->seqcount irq_context: 0 &type->s_umount_key#18/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#18/1 sb_lock irq_context: 0 &type->s_umount_key#18/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#18/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 irq_context: 0 &type->s_umount_key#18/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#18/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &type->s_umount_key#18/1 &dentry->d_lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &____s->seqcount irq_context: 0 tomoyo_ss irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 tomoyo_ss (console_sem).lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner_lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner irq_context: 0 tomoyo_ss console_lock console_srcu console_owner &port_lock_key irq_context: 0 tomoyo_ss console_lock console_srcu console_owner console_owner_lock irq_context: 0 pnp_lock irq_context: 0 pnp_lock fs_reclaim irq_context: 0 pnp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pnp_lock pool_lock#2 irq_context: 0 &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 &device->physical_node_lock fs_reclaim irq_context: 0 &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &device->physical_node_lock pool_lock#2 irq_context: 0 &device->physical_node_lock lock irq_context: 0 &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 fwnode_link_lock irq_context: 0 fwnode_link_lock &k->k_lock irq_context: 0 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex device_links_lock irq_context: 0 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex lock irq_context: 0 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex pnp_lock irq_context: 0 &dev->mutex resource_lock irq_context: 0 &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex fwnode_link_lock &k->k_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex probe_waitqueue.lock irq_context: 0 subsys mutex#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem lock#4 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_lock_key#22 irq_context: 0 jiffies_seq.seqcount irq_context: 0 subsys mutex#23 irq_context: 0 subsys mutex#23 &k->k_lock irq_context: 0 subsys mutex#16 irq_context: 0 subsys mutex#16 &k->k_lock irq_context: 0 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override pool_lock irq_context: 0 subsys mutex#24 irq_context: 0 subsys mutex#24 &k->k_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u8.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 kfence_freelist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &meta->lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override &c->lock irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_callback pcpu_lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: softirq led_lock irq_context: 0 misc_mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 subsys mutex#25 irq_context: 0 subsys mutex#25 &k->list_lock irq_context: 0 subsys mutex#25 &k->k_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &xa->xa_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex kthread_create_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &x->wait irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 netevent_notif_chain.lock irq_context: 0 clients_rwsem irq_context: 0 clients_rwsem fs_reclaim irq_context: 0 clients_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 clients_rwsem clients.xa_lock irq_context: 0 devices_rwsem irq_context: 0 clients_rwsem clients.xa_lock pool_lock#2 irq_context: 0 (blocking_lsm_notifier_chain).rwsem irq_context: 0 (inetaddr_chain).rwsem irq_context: 0 inet6addr_chain.lock irq_context: 0 buses_mutex irq_context: 0 offload_lock irq_context: 0 inetsw_lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events pcpu_balance_work irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex free_vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 ptype_lock irq_context: 0 (wq_completion)events_power_efficient irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem slab_mutex irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem slab_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem slab_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 tcp_ulp_list_lock irq_context: 0 xfrm_state_afinfo_lock irq_context: 0 xfrm_policy_afinfo_lock irq_context: 0 xfrm_input_afinfo_lock irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock irq_context: 0 rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock &base->lock irq_context: 0 rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET/1 irq_context: 0 (wq_completion)events_highpri irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &____s->seqcount irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) krc.lock irq_context: 0 &hashinfo->lock irq_context: 0 k-slock-AF_INET/1 irq_context: 0 tcp_cong_list_lock irq_context: 0 rcu_read_lock &____s->seqcount irq_context: 0 mptcp_sched_list_lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) cache_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock &obj_hash[i].lock irq_context: 0 (rpc_pipefs_notifier_list).rwsem irq_context: 0 svc_xprt_class_lock irq_context: 0 xprt_list_lock irq_context: 0 xprt_list_lock (console_sem).lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner_lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner irq_context: 0 xprt_list_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 xprt_list_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 umhelper_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock init_fs.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 mount_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_mm.page_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_files.file_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &rq->__lock irq_context: 0 umh_sysctl_lock irq_context: 0 async_done.lock irq_context: 0 &drv->dynids.lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &dentry->d_lock pool_lock#2 irq_context: 0 &tsk->futex_exit_mutex irq_context: 0 &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &p->alloc_lock &fs->lock irq_context: 0 &child->perf_event_mutex irq_context: 0 css_set_lock irq_context: 0 tasklist_lock irq_context: 0 tasklist_lock &pid->wait_pidfd irq_context: 0 tasklist_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit irq_context: 0 tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pool_lock#2 irq_context: 0 tasklist_lock &obj_hash[i].lock irq_context: 0 &cfs_rq->removed.lock irq_context: 0 misc_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 subsys mutex#26 irq_context: 0 subsys mutex#27 irq_context: 0 subsys mutex#27 &k->list_lock irq_context: 0 subsys mutex#27 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock pool_lock#2 irq_context: 0 subsys mutex#28 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pmus_lock fs_reclaim irq_context: 0 pmus_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pmus_lock &k->list_lock irq_context: 0 pmus_lock lock irq_context: 0 pmus_lock lock kernfs_idr_lock irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pmus_lock uevent_sock_mutex irq_context: 0 pmus_lock rcu_read_lock &pool->lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pmus_lock running_helpers_waitq.lock irq_context: 0 pmus_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 pmus_lock &c->lock irq_context: 0 pmus_lock &pcp->lock &zone->lock irq_context: 0 pmus_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pmus_lock &____s->seqcount irq_context: 0 pmus_lock &x->wait#9 irq_context: 0 pmus_lock bus_type_sem irq_context: 0 pmus_lock sysfs_symlink_target_lock irq_context: 0 pmus_lock &k->k_lock irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &dev->power.lock irq_context: 0 pmus_lock dpm_list_mtx irq_context: 0 pmus_lock &dev->mutex &k->list_lock irq_context: 0 pmus_lock &dev->mutex &k->k_lock irq_context: 0 pmus_lock &dev->mutex &dev->power.lock irq_context: 0 pmus_lock subsys mutex#29 irq_context: 0 pmus_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pmus_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pmus_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 key_user_lock irq_context: 0 key_serial_lock irq_context: 0 key_construction_mutex irq_context: 0 &type->lock_class irq_context: 0 &type->lock_class keyring_serialise_link_lock irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 keyring_serialise_link_lock irq_context: 0 &pgdat->kswapd_lock fs_reclaim irq_context: 0 &pgdat->kswapd_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pgdat->kswapd_lock pool_lock#2 irq_context: 0 &pgdat->kswapd_lock kthread_create_lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kswapd_lock &x->wait irq_context: 0 &pgdat->kswapd_lock &rq->__lock irq_context: 0 &pgdat->kswapd_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kswapd_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &rq->__lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_wait irq_context: 0 list_lrus_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex pool_lock#2 irq_context: 0 drivers_lock irq_context: 0 damon_dbgfs_lock irq_context: 0 damon_dbgfs_lock fs_reclaim irq_context: 0 damon_dbgfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock pool_lock#2 irq_context: 0 damon_dbgfs_lock damon_ops_lock irq_context: 0 damon_dbgfs_lock pin_fs_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 misc_mtx &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#19/1 irq_context: 0 &type->s_umount_key#19/1 fs_reclaim irq_context: 0 &type->s_umount_key#19/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 pool_lock#2 irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#19/1 shrinker_mutex irq_context: 0 &type->s_umount_key#19/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#19/1 sb_lock irq_context: 0 &type->s_umount_key#19/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#19/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 irq_context: 0 &type->s_umount_key#19/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#19/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#19/1 &c->lock irq_context: 0 &type->s_umount_key#19/1 &____s->seqcount irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 &dentry->d_lock irq_context: 0 &type->s_umount_key#19/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#20/1 irq_context: 0 &type->s_umount_key#20/1 fs_reclaim irq_context: 0 &type->s_umount_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#20/1 pool_lock#2 irq_context: 0 &type->s_umount_key#20/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#20/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#20/1 &c->lock irq_context: 0 &type->s_umount_key#20/1 &____s->seqcount irq_context: 0 &type->s_umount_key#20/1 shrinker_mutex irq_context: 0 &type->s_umount_key#20/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#20/1 sb_lock irq_context: 0 &type->s_umount_key#20/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#20/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#20/1 &sb->s_type->i_lock_key#18 irq_context: 0 &type->s_umount_key#20/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#20/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#20/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &type->s_umount_key#20/1 &dentry->d_lock irq_context: 0 configfs_subsystem_mutex irq_context: 0 &sb->s_type->i_mutex_key#6/1 irq_context: 0 &sb->s_type->i_mutex_key#6/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 irq_context: 0 misc_mtx rcu_read_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slab_mutex rcu_read_lock &pool->lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ecryptfs_daemon_hash_mux irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ecryptfs_daemon_hash_mux pool_lock#2 irq_context: 0 &ecryptfs_kthread_ctl.wait irq_context: 0 ecryptfs_msg_ctx_lists_mux irq_context: 0 ecryptfs_msg_ctx_lists_mux &ecryptfs_msg_ctx_arr[i].mux irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem running_helpers_waitq.lock irq_context: 0 nfs_version_lock irq_context: 0 key_types_sem irq_context: 0 key_types_sem (console_sem).lock irq_context: 0 key_types_sem console_lock console_srcu console_owner_lock irq_context: 0 key_types_sem console_lock console_srcu console_owner irq_context: 0 key_types_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 key_types_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 pnfs_spinlock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem krc.lock irq_context: 0 pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem kfence_freelist_lock irq_context: 0 nls_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&cache_cleaner)->timer irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback put_task_map-wait-type-override quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) irq_context: 0 (wq_completion)events (work_completion)(&p->wq) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) pool_lock#2 irq_context: 0 jffs2_compressor_list_lock irq_context: 0 misc_mtx &pcp->lock &zone->lock irq_context: 0 misc_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock irq_context: 0 next_tag_value_lock irq_context: softirq (&tcp_orphan_timer) irq_context: softirq (&tcp_orphan_timer) &obj_hash[i].lock irq_context: softirq (&tcp_orphan_timer) &base->lock irq_context: softirq (&tcp_orphan_timer) &base->lock &obj_hash[i].lock irq_context: 0 log_redrive_lock irq_context: 0 &TxAnchor.LazyLock irq_context: 0 &TxAnchor.LazyLock jfs_commit_thread_wait.lock irq_context: 0 jfsTxnLock irq_context: 0 ocfs2_stack_lock irq_context: 0 ocfs2_stack_lock (console_sem).lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner_lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 o2hb_callback_sem irq_context: 0 o2net_handler_lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#30 irq_context: 0 subsys mutex#30 &k->k_lock irq_context: 0 &type->s_umount_key#21/1 irq_context: 0 &type->s_umount_key#21/1 fs_reclaim irq_context: 0 &type->s_umount_key#21/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#21/1 shrinker_mutex irq_context: 0 &type->s_umount_key#21/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#21/1 sb_lock irq_context: 0 &type->s_umount_key#21/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#21/1 &____s->seqcount irq_context: 0 &type->s_umount_key#21/1 &c->lock irq_context: 0 &type->s_umount_key#21/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#19 irq_context: 0 &type->s_umount_key#21/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#21/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &type->s_umount_key#21/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 irq_context: 0 &type->s_umount_key#22/1 fs_reclaim irq_context: 0 &type->s_umount_key#22/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 pool_lock#2 irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#22/1 shrinker_mutex irq_context: 0 &type->s_umount_key#22/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#22/1 sb_lock irq_context: 0 &type->s_umount_key#22/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#22/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#20 irq_context: 0 &type->s_umount_key#22/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#22/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#20 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 &c->lock irq_context: 0 &type->s_umount_key#22/1 &____s->seqcount irq_context: 0 cipso_v4_doi_list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 crypto_alg_sem irq_context: 0 alg_types_sem irq_context: 0 alg_types_sem fs_reclaim irq_context: 0 alg_types_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 alg_types_sem pool_lock#2 irq_context: 0 dma_list_mutex irq_context: 0 asymmetric_key_parsers_sem irq_context: 0 asymmetric_key_parsers_sem (console_sem).lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner_lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 blkcg_pol_register_mutex irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex cgroup_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock kernfs_idr_lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 elv_list_lock irq_context: 0 crc_t10dif_mutex irq_context: 0 crc_t10dif_mutex crypto_alg_sem irq_context: 0 crc_t10dif_mutex fs_reclaim irq_context: 0 crc_t10dif_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc_t10dif_mutex pool_lock#2 irq_context: 0 crc64_rocksoft_mutex irq_context: 0 crc64_rocksoft_mutex crypto_alg_sem irq_context: 0 crc64_rocksoft_mutex fs_reclaim irq_context: 0 crc64_rocksoft_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc64_rocksoft_mutex pool_lock#2 irq_context: 0 ts_mod_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key &c->lock irq_context: 0 &sb->s_type->i_mutex_key &____s->seqcount irq_context: 0 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &____s->seqcount irq_context: 0 pci_ep_cfs_subsys.su_mutex irq_context: 0 &default_group_class[depth - 1]#2/1 irq_context: 0 &default_group_class[depth - 1]#2/1 fs_reclaim irq_context: 0 &default_group_class[depth - 1]#2/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &default_group_class[depth - 1]#2/1 pool_lock#2 irq_context: 0 &default_group_class[depth - 1]#2/1 &dentry->d_lock irq_context: 0 &default_group_class[depth - 1]#2/1 configfs_dirent_lock irq_context: 0 &default_group_class[depth - 1]#2/1 mmu_notifier_invalidate_range_start irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_lock_key#18 irq_context: 0 &default_group_class[depth - 1]#2/1 &s->s_inode_list_lock irq_context: 0 &default_group_class[depth - 1]#2/1 tk_core.seq.seqcount irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_mutex_key#7/2 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_mutex_key#7/2 irq_context: 0 pci_epf_mutex irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 ipmi_interfaces_mutex irq_context: 0 ipmi_interfaces_mutex &k->list_lock irq_context: 0 ipmi_interfaces_mutex fs_reclaim irq_context: 0 ipmi_interfaces_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ipmi_interfaces_mutex pool_lock#2 irq_context: 0 ipmi_interfaces_mutex lock irq_context: 0 ipmi_interfaces_mutex lock kernfs_idr_lock irq_context: 0 ipmi_interfaces_mutex &root->kernfs_rwsem irq_context: 0 ipmi_interfaces_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 ipmi_interfaces_mutex &k->k_lock irq_context: 0 ipmi_interfaces_mutex &c->lock irq_context: 0 ipmi_interfaces_mutex &____s->seqcount irq_context: 0 ipmi_interfaces_mutex uevent_sock_mutex irq_context: 0 ipmi_interfaces_mutex &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 ipmi_interfaces_mutex running_helpers_waitq.lock irq_context: softirq rcu_callback &meta->lock irq_context: softirq rcu_callback kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock pool_lock irq_context: 0 tasklist_lock quarantine_lock irq_context: 0 ipmi_interfaces_mutex &obj_hash[i].lock pool_lock irq_context: 0 ipmi_interfaces_mutex pcpu_alloc_mutex irq_context: 0 ipmi_interfaces_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex kthread_create_lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock &rq->__lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ipmi_interfaces_mutex &rq->__lock irq_context: 0 ipmi_interfaces_mutex &x->wait irq_context: 0 ipmi_interfaces_mutex wq_pool_mutex irq_context: 0 ipmi_interfaces_mutex wq_pool_mutex &wq->mutex irq_context: 0 ipmi_interfaces_mutex &base->lock irq_context: 0 ipmi_interfaces_mutex &base->lock &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex panic_notifier_list.lock irq_context: 0 smi_watchers_mutex irq_context: 0 smi_watchers_mutex &ipmi_interfaces_srcu irq_context: 0 smi_infos_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &device->physical_node_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 &dev->mutex acpi_wakeup_lock irq_context: 0 &dev->mutex semaphore->lock irq_context: 0 &dev->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex irq_domain_mutex irq_context: 0 &dev->mutex &domain->mutex irq_context: 0 &dev->mutex kthread_create_lock irq_context: 0 &dev->mutex &p->pi_lock irq_context: 0 &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &x->wait irq_context: 0 &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &desc->request_mutex irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &desc->wait_for_threads irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &p->pi_lock &rq->__lock &rt_rq->rt_runtime_lock irq_context: 0 &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &dev->mutex register_lock fs_reclaim irq_context: 0 &dev->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_lock pool_lock#2 irq_context: 0 &dev->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &dev->mutex register_lock &c->lock irq_context: 0 &dev->mutex register_lock &____s->seqcount irq_context: 0 &dev->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex gdp_mutex irq_context: 0 &dev->mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex gdp_mutex lock irq_context: 0 &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex bus_type_sem irq_context: 0 &dev->mutex &c->lock irq_context: 0 &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#31 irq_context: 0 &dev->mutex subsys mutex#31 &k->k_lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock irq_context: 0 &dev->mutex input_mutex irq_context: 0 &dev->mutex input_mutex fs_reclaim irq_context: 0 &dev->mutex input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex input_mutex pool_lock#2 irq_context: 0 &dev->mutex input_mutex &dev->mutex#2 irq_context: 0 &dev->mutex input_mutex input_devices_poll_wait.lock irq_context: 0 &dev->mutex wakeup_ida.xa_lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#13 irq_context: 0 &dev->mutex subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex events_lock irq_context: 0 register_count_mutex irq_context: 0 register_count_mutex &k->list_lock irq_context: 0 register_count_mutex fs_reclaim irq_context: 0 register_count_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_count_mutex pool_lock#2 irq_context: 0 register_count_mutex lock irq_context: 0 register_count_mutex lock kernfs_idr_lock irq_context: 0 register_count_mutex &root->kernfs_rwsem irq_context: 0 register_count_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_count_mutex &k->k_lock irq_context: 0 register_count_mutex &c->lock irq_context: 0 register_count_mutex &____s->seqcount irq_context: 0 register_count_mutex uevent_sock_mutex irq_context: 0 register_count_mutex &obj_hash[i].lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_count_mutex running_helpers_waitq.lock irq_context: 0 register_count_mutex &rq->__lock irq_context: 0 &dev->mutex cpu_add_remove_lock irq_context: 0 &dev->mutex thermal_cdev_ida.xa_lock irq_context: 0 &dev->mutex cpufreq_driver_lock irq_context: 0 &dev->mutex subsys mutex#32 irq_context: 0 &dev->mutex subsys mutex#32 &k->k_lock irq_context: 0 &dev->mutex thermal_list_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock pool_lock irq_context: 0 scmi_requested_devices_mtx irq_context: 0 scmi_requested_devices_mtx fs_reclaim irq_context: 0 scmi_requested_devices_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 scmi_requested_devices_mtx pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &dev->mutex iommu_probe_device_lock irq_context: 0 &dev->mutex acpi_link_lock irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_link_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_link_lock semaphore->lock irq_context: 0 &dev->mutex acpi_link_lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex acpi_link_lock &____s->seqcount irq_context: 0 &dev->mutex acpi_link_lock (console_sem).lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex &drv->dynids.lock irq_context: 0 &dev->mutex pci_lock irq_context: 0 &dev->mutex virtio_index_ida.xa_lock irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#33 irq_context: 0 vdpa_dev_lock irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock &wq irq_context: 0 subsys mutex#34 irq_context: 0 subsys mutex#34 &k->k_lock irq_context: 0 rcu_read_lock rcu_node_0 irq_context: 0 sched_map-wait-type-override &rq->__lock irq_context: hardirq|softirq &dev->power.lock rcu_read_lock &pool->lock irq_context: hardirq|softirq &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq|softirq &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: hardirq|softirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq lib/debugobjects.c:101 irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (debug_obj_work).work irq_context: softirq rcu_callback &base->lock irq_context: softirq rcu_callback &base->lock &obj_hash[i].lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 quarantine_lock irq_context: 0 (wq_completion)events (debug_obj_work).work pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &____s->seqcount irq_context: 0 port_mutex irq_context: 0 port_mutex fs_reclaim irq_context: 0 port_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex pool_lock#2 irq_context: 0 port_mutex &x->wait#9 irq_context: 0 port_mutex &obj_hash[i].lock irq_context: 0 port_mutex &obj_hash[i].lock pool_lock irq_context: 0 port_mutex &k->list_lock irq_context: 0 port_mutex lock irq_context: 0 port_mutex lock kernfs_idr_lock irq_context: 0 port_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex bus_type_sem irq_context: 0 port_mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &c->lock irq_context: 0 port_mutex &____s->seqcount irq_context: 0 port_mutex &k->k_lock irq_context: 0 port_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &dev->power.lock irq_context: 0 port_mutex dpm_list_mtx irq_context: 0 port_mutex uevent_sock_mutex irq_context: 0 port_mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex running_helpers_waitq.lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock irq_context: 0 port_mutex &dev->mutex &k->list_lock irq_context: 0 port_mutex &dev->mutex &k->k_lock irq_context: 0 port_mutex &dev->mutex device_links_srcu irq_context: 0 port_mutex &dev->mutex fwnode_link_lock irq_context: 0 port_mutex &dev->mutex device_links_lock irq_context: 0 port_mutex &dev->mutex fs_reclaim irq_context: 0 port_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &dev->mutex pool_lock#2 irq_context: 0 port_mutex &dev->mutex &dev->devres_lock irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex irq_context: 0 port_mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 port_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &dev->mutex lock irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 port_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &dev->mutex deferred_probe_mutex irq_context: 0 port_mutex &dev->mutex &c->lock irq_context: 0 port_mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 port_mutex &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 port_mutex &dev->mutex &____s->seqcount irq_context: 0 port_mutex &dev->mutex uevent_sock_mutex irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 port_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 port_mutex subsys mutex#14 irq_context: 0 port_mutex &xa->xa_lock#3 irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &dev->mutex &rq->__lock irq_context: 0 port_mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 port_mutex &port->mutex irq_context: 0 port_mutex &port->mutex fs_reclaim irq_context: 0 port_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex pool_lock#2 irq_context: 0 port_mutex &port->mutex console_mutex irq_context: 0 port_mutex &port->mutex ctrl_ida.xa_lock irq_context: 0 port_mutex &port->mutex &x->wait#9 irq_context: 0 port_mutex &port->mutex &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex &dev->power.lock irq_context: 0 port_mutex &port->mutex &k->list_lock irq_context: 0 port_mutex &port->mutex lock irq_context: 0 port_mutex &port->mutex lock kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &port->mutex bus_type_sem irq_context: 0 port_mutex &port->mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &port->mutex &k->k_lock irq_context: 0 port_mutex &port->mutex dpm_list_mtx irq_context: 0 port_mutex &port->mutex uevent_sock_mutex irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex running_helpers_waitq.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 port_mutex &port->mutex &dev->mutex &k->list_lock irq_context: 0 port_mutex &port->mutex &dev->mutex &k->k_lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex subsys mutex#15 irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 port_mutex &port->mutex kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &k->k_lock klist_remove_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 port_mutex &port->mutex deferred_probe_mutex irq_context: 0 port_mutex &port->mutex device_links_lock irq_context: 0 port_mutex &port->mutex mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex &c->lock irq_context: 0 port_mutex &port->mutex &____s->seqcount irq_context: 0 port_mutex &port->mutex gdp_mutex irq_context: 0 port_mutex &port->mutex gdp_mutex &k->list_lock irq_context: 0 port_mutex &port->mutex gdp_mutex fs_reclaim irq_context: 0 port_mutex &port->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex gdp_mutex pool_lock#2 irq_context: 0 port_mutex &port->mutex gdp_mutex lock irq_context: 0 port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &port->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 port_mutex &port->mutex req_lock irq_context: 0 port_mutex &port->mutex &p->pi_lock irq_context: 0 port_mutex &port->mutex &p->pi_lock &rq->__lock irq_context: 0 port_mutex &port->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 port_mutex &port->mutex &rq->__lock irq_context: 0 port_mutex &port->mutex &x->wait#11 irq_context: 0 port_mutex &port->mutex subsys mutex#16 irq_context: 0 port_mutex &port->mutex subsys mutex#16 &k->k_lock irq_context: 0 port_mutex &port->mutex chrdevs_lock irq_context: 0 port_mutex &port->mutex &obj_hash[i].lock pool_lock irq_context: 0 port_mutex &port->mutex &cfs_rq->removed.lock irq_context: 0 port_mutex &port->mutex quarantine_lock irq_context: 0 port_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (work_completion)(&buf->work) irq_context: 0 uevent_sock_mutex &rq->__lock irq_context: 0 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 uevent_sock_mutex pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex rng_index_ida.xa_lock irq_context: 0 &dev->mutex &md->mutex irq_context: 0 &dev->mutex free_vmap_area_lock irq_context: 0 &dev->mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &dev->mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &dev->mutex vmap_area_lock irq_context: 0 &dev->mutex &md->mutex pci_lock irq_context: 0 &dev->mutex &md->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#6 irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#6 pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &rq->__lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex tmpmask_lock irq_context: 0 &dev->mutex &md->mutex &its->lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex lock irq_context: 0 &dev->mutex &md->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock batched_entropy_u8.lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock kfence_freelist_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 &dev->mutex &zone->lock irq_context: 0 &dev->mutex &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &dev->vqs_list_lock irq_context: 0 &dev->mutex &vp_dev->lock irq_context: 0 &dev->mutex rng_mutex irq_context: 0 &dev->mutex rng_mutex &x->wait#13 irq_context: 0 &dev->mutex rng_mutex fs_reclaim irq_context: 0 &dev->mutex rng_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rng_mutex pool_lock#2 irq_context: 0 &dev->mutex rng_mutex kthread_create_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rng_mutex &rq->__lock irq_context: 0 &dev->mutex rng_mutex &x->wait irq_context: hardirq &x->wait#14 irq_context: hardirq &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: hardirq &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex rng_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex reading_mutex irq_context: 0 rng_mutex irq_context: 0 &dev->mutex input_pool.lock irq_context: 0 reading_mutex irq_context: 0 &dev->mutex &dev->config_lock irq_context: 0 reading_mutex &x->wait#14 irq_context: 0 misc_mtx &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq drivers/char/random.c:1010 irq_context: softirq drivers/char/random.c:1010 input_pool.lock irq_context: 0 &dev->devres_lock irq_context: 0 &dev->managed.lock irq_context: 0 &type->s_umount_key#23/1 irq_context: 0 &type->s_umount_key#23/1 fs_reclaim irq_context: 0 &type->s_umount_key#23/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#23/1 shrinker_mutex irq_context: 0 &type->s_umount_key#23/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#23/1 sb_lock irq_context: 0 &type->s_umount_key#23/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#23/1 &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#21 irq_context: 0 &type->s_umount_key#23/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#23/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#21 &dentry->d_lock irq_context: 0 &type->s_umount_key#23/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#21 irq_context: 0 lock drm_minor_lock irq_context: 0 lock drm_minor_lock pool_lock#2 irq_context: 0 stack_depot_init_mutex irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#35 irq_context: 0 subsys mutex#35 &k->k_lock irq_context: 0 drm_minor_lock irq_context: 0 &dev->mode_config.idr_mutex irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mode_config.idr_mutex pool_lock#2 irq_context: 0 crtc_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &dev->mode_config.blob_lock irq_context: 0 &xa->xa_lock#7 irq_context: 0 &xa->xa_lock#8 irq_context: 0 &dev->mode_config.connector_list_lock irq_context: 0 &dev->vbl_lock irq_context: 0 drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 drm_connector_list_iter fs_reclaim irq_context: 0 drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &x->wait#9 irq_context: 0 drm_connector_list_iter &connector->mutex &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->list_lock irq_context: 0 drm_connector_list_iter &connector->mutex lock irq_context: 0 drm_connector_list_iter &connector->mutex lock kernfs_idr_lock irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex bus_type_sem irq_context: 0 drm_connector_list_iter &connector->mutex sysfs_symlink_target_lock irq_context: 0 drm_connector_list_iter &connector->mutex &c->lock irq_context: 0 drm_connector_list_iter &connector->mutex &____s->seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &dev->power.lock irq_context: 0 drm_connector_list_iter &connector->mutex dpm_list_mtx irq_context: 0 drm_connector_list_iter &connector->mutex uevent_sock_mutex irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 drm_connector_list_iter &connector->mutex running_helpers_waitq.lock irq_context: 0 drm_connector_list_iter &connector->mutex &rq->__lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#35 irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#35 &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex pin_fs_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &dev->mode_config.idr_mutex irq_context: 0 drm_connector_list_iter &connector->mutex connector_list_lock irq_context: 0 &dev->filelist_mutex irq_context: 0 &dev->clientlist_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &sbinfo->stat_lock irq_context: 0 &dev->clientlist_mutex &helper->lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex &helper->lock &s->s_inode_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock batched_entropy_u32.lock irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock &c->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.fb_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file->fbs_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &prime_fpriv->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &c->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock req_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#11 irq_context: 0 &dev->clientlist_mutex registration_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#11 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->clientlist_mutex registration_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->clientlist_mutex registration_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &fb_info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock vt_event_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock console_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#6 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#6 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.blob_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &crtc->commit_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &xa->xa_lock#9 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex lock#4 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &xa->xa_lock#9 pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex lock#4 &lruvec->lru_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &____s->seqcount#5 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &x->wait#15 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#5 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &x->wait#15 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->damage_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->damage_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock reservation_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vkms_out->lock irq_context: hardirq &vkms_out->lock &dev->event_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#5 irq_context: hardirq &vkms_out->lock &dev->event_lock &vblank->queue irq_context: hardirq &vkms_out->lock &dev->event_lock &____s->seqcount#5 irq_context: hardirq &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock &rq->__lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (&timer.timer) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work)#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_owner_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &lock->wait_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex (console_sem).lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex kernel_fb_helper_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock irq_context: 0 drivers_lock#2 irq_context: 0 devices_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#9 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up bus_type_sem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->k_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock irq_context: 0 blk_queue_ida.xa_lock irq_context: 0 &sb->s_type->i_lock_key#3 irq_context: 0 &xa->xa_lock#10 irq_context: 0 lock &q->queue_lock irq_context: 0 lock &q->queue_lock &blkcg->lock irq_context: 0 &q->queue_lock irq_context: 0 &q->queue_lock pool_lock#2 irq_context: 0 &q->queue_lock pcpu_lock irq_context: 0 &q->queue_lock &obj_hash[i].lock irq_context: 0 &q->queue_lock percpu_counters_lock irq_context: 0 &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->queue_lock &blkcg->lock irq_context: 0 &bdev->bd_size_lock irq_context: 0 subsys mutex#36 irq_context: 0 subsys mutex#36 &k->k_lock irq_context: 0 dev_hotplug_mutex irq_context: 0 dev_hotplug_mutex &dev->power.lock irq_context: 0 &q->sysfs_dir_lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 percpu_ref_switch_lock irq_context: 0 subsys mutex#37 irq_context: 0 subsys mutex#37 &k->k_lock irq_context: 0 cgwb_lock irq_context: 0 bdi_lock irq_context: 0 inode_hash_lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 bdev_lock irq_context: 0 &disk->open_mutex irq_context: 0 &disk->open_mutex fs_reclaim irq_context: 0 &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex pool_lock#2 irq_context: 0 &disk->open_mutex free_vmap_area_lock irq_context: 0 &disk->open_mutex vmap_area_lock irq_context: 0 &disk->open_mutex &____s->seqcount irq_context: 0 &disk->open_mutex init_mm.page_table_lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 irq_context: 0 &disk->open_mutex lock#4 irq_context: 0 &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &disk->open_mutex &c->lock irq_context: 0 &disk->open_mutex &mapping->i_private_lock irq_context: 0 &disk->open_mutex tk_core.seq.seqcount irq_context: 0 &disk->open_mutex &ret->b_uptodate_lock irq_context: 0 &disk->open_mutex &obj_hash[i].lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 pool_lock#2 irq_context: 0 &disk->open_mutex purge_vmap_area_lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &disk->open_mutex lock#4 &lruvec->lru_lock irq_context: 0 &disk->open_mutex lock#5 irq_context: 0 &disk->open_mutex &lruvec->lru_lock irq_context: 0 lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &q->queue_lock &c->lock irq_context: 0 &q->queue_lock &____s->seqcount irq_context: 0 &disk->open_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock pool_lock#2 irq_context: 0 &disk->open_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock pool_lock#2 irq_context: 0 loop_ctl_mutex irq_context: 0 loop_ctl_mutex fs_reclaim irq_context: 0 loop_ctl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 loop_ctl_mutex pool_lock#2 irq_context: 0 &q->sysfs_lock irq_context: 0 &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_lock cpu_hotplug_lock irq_context: 0 &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock &xa->xa_lock#11 irq_context: 0 &set->tag_list_lock irq_context: 0 &q->mq_freeze_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->queue_lock &pcp->lock &zone->lock irq_context: 0 &q->queue_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pcpu_alloc_mutex &rq->__lock irq_context: 0 pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 pcpu_alloc_mutex pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &c->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 rcu_read_lock &c->lock irq_context: 0 nbd_index_mutex irq_context: 0 nbd_index_mutex fs_reclaim irq_context: 0 nbd_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nbd_index_mutex pool_lock#2 irq_context: 0 set->srcu irq_context: 0 (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (work_completion)(&(&hctx->run_work)->work) irq_context: 0 &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &q->sysfs_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->sysfs_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->sysfs_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &q->sysfs_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock crngs.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: softirq &(&ops->cursor_work)->timer irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) (console_sem).lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock &helper->damage_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 zram_index_mutex irq_context: 0 zram_index_mutex fs_reclaim irq_context: 0 zram_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex pool_lock#2 irq_context: 0 zram_index_mutex blk_queue_ida.xa_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock irq_context: 0 zram_index_mutex pcpu_alloc_mutex irq_context: 0 zram_index_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 zram_index_mutex bio_slab_lock irq_context: 0 zram_index_mutex &c->lock irq_context: 0 zram_index_mutex &____s->seqcount irq_context: 0 zram_index_mutex percpu_counters_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock pool_lock irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 zram_index_mutex mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex &s->s_inode_list_lock irq_context: 0 zram_index_mutex &xa->xa_lock#10 irq_context: 0 zram_index_mutex lock irq_context: 0 zram_index_mutex lock &q->queue_lock irq_context: 0 zram_index_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 zram_index_mutex &q->queue_lock irq_context: 0 zram_index_mutex &q->queue_lock &c->lock irq_context: 0 zram_index_mutex &q->queue_lock &____s->seqcount irq_context: 0 zram_index_mutex &q->queue_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->queue_lock pcpu_lock irq_context: 0 zram_index_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 zram_index_mutex &q->queue_lock percpu_counters_lock irq_context: 0 zram_index_mutex &q->queue_lock &blkcg->lock irq_context: 0 zram_index_mutex &x->wait#9 irq_context: 0 zram_index_mutex &bdev->bd_size_lock irq_context: 0 zram_index_mutex &k->list_lock irq_context: 0 zram_index_mutex gdp_mutex irq_context: 0 zram_index_mutex gdp_mutex &k->list_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex bus_type_sem irq_context: 0 zram_index_mutex sysfs_symlink_target_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &dev->power.lock irq_context: 0 zram_index_mutex dpm_list_mtx irq_context: 0 zram_index_mutex req_lock irq_context: 0 zram_index_mutex &p->pi_lock irq_context: 0 zram_index_mutex &x->wait#11 irq_context: 0 zram_index_mutex &rq->__lock irq_context: 0 zram_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 zram_index_mutex subsys mutex#36 irq_context: 0 zram_index_mutex subsys mutex#36 &k->k_lock irq_context: 0 zram_index_mutex dev_hotplug_mutex irq_context: 0 zram_index_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex percpu_ref_switch_lock irq_context: 0 zram_index_mutex uevent_sock_mutex irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 zram_index_mutex running_helpers_waitq.lock irq_context: 0 zram_index_mutex subsys mutex#37 irq_context: 0 zram_index_mutex subsys mutex#37 &k->k_lock irq_context: 0 zram_index_mutex cgwb_lock irq_context: 0 zram_index_mutex pin_fs_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex bdi_lock irq_context: 0 zram_index_mutex inode_hash_lock irq_context: 0 zram_index_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex (console_sem).lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner_lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner irq_context: 0 zram_index_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 zram_index_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 subsys mutex#38 irq_context: 0 subsys mutex#38 &k->k_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 configfs_dirent_lock irq_context: 0 &q->sysfs_lock &xa->xa_lock#11 pool_lock#2 irq_context: 0 &lock irq_context: 0 &lock nullb_indexes.xa_lock irq_context: 0 &q->sysfs_dir_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &ret->b_uptodate_lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_node_0 irq_context: 0 nfc_index_ida.xa_lock irq_context: 0 nfc_devlist_mutex irq_context: 0 nfc_devlist_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex irq_context: 0 nfc_devlist_mutex gdp_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex gdp_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex gdp_mutex lock irq_context: 0 nfc_devlist_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex lock irq_context: 0 nfc_devlist_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex bus_type_sem irq_context: 0 nfc_devlist_mutex sysfs_symlink_target_lock irq_context: 0 nfc_devlist_mutex &c->lock irq_context: 0 nfc_devlist_mutex &____s->seqcount irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &dev->power.lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx irq_context: 0 nfc_devlist_mutex uevent_sock_mutex irq_context: 0 nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 nfc_devlist_mutex running_helpers_waitq.lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 irq_context: 0 nfc_devlist_mutex subsys mutex#39 &k->k_lock irq_context: 0 llcp_devices_lock irq_context: 0 &dev->mutex rfkill_global_mutex irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex &k->list_lock irq_context: 0 &dev->mutex rfkill_global_mutex lock irq_context: 0 &dev->mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex bus_type_sem irq_context: 0 &dev->mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rfkill_global_mutex &____s->seqcount irq_context: 0 &dev->mutex rfkill_global_mutex &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &dev->mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &dev->mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &dev->mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rfkill_global_mutex &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex leds_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rfkill->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dma_heap_minors.xa_lock irq_context: 0 subsys mutex#41 irq_context: 0 subsys mutex#41 &k->k_lock irq_context: 0 heap_list_lock irq_context: 0 dma_heap_minors.xa_lock pool_lock#2 irq_context: 0 misc_mtx lock kernfs_idr_lock &c->lock irq_context: 0 misc_mtx lock kernfs_idr_lock &____s->seqcount irq_context: 0 subsys mutex#42 irq_context: 0 subsys mutex#42 &k->list_lock irq_context: 0 subsys mutex#42 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex semaphore->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex iommu_probe_device_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 nvmf_hosts_mutex irq_context: 0 subsys mutex#43 irq_context: 0 subsys mutex#43 &k->k_lock irq_context: 0 nvmf_transports_rwsem irq_context: 0 subsys mutex#44 irq_context: 0 subsys mutex#44 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock semaphore->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex irq_domain_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &drv->dynids.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cpu_add_remove_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_instance_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex chrdevs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#45 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#45 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx pm_qos_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex resource_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock pools_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &xa->xa_lock#6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &xa->xa_lock#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &set->tag_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->lock &ctrl->state_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock &obj_hash[i].lock irq_context: softirq &x->wait#16 irq_context: softirq &x->wait#16 &p->pi_lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#16 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock subsys mutex#46 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock subsys mutex#46 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &xa->xa_lock#12 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class irq_resend_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &ent->pde_unload_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &xa->xa_lock#6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &xa->xa_lock#6 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &cma->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->namespaces_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (work_completion)(&ctrl->scan_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock (wq_completion)nvme-wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#10 irq_context: 0 (wq_completion)nvme-wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &hctx->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &hctx->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#16 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &base->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#18 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 irq_context: 0 nvmet_config_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock (&timer.timer) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &____s->seqcount irq_context: 0 subsys mutex#47 irq_context: 0 subsys mutex#47 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->async_event_work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->async_event_work) &nvmeq->sq_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &ctrl->namespaces_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock blk_queue_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &xa->xa_lock#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &xa->xa_lock#11 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &set->tag_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bio_slab_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock percpu_counters_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &xa->xa_lock#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock &q->queue_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock percpu_counters_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock nvme_subsystems_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &xa->xa_lock#12 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &bdev->bd_size_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &ctrl->namespaces_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock ext_devt_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &k->list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bus_type_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dpm_list_mtx irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock req_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#36 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#36 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dev_hotplug_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7/2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 backend_mutex irq_context: 0 scsi_mib_index_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 hba_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 device_mutex irq_context: 0 device_mutex fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 device_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 device_mutex pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock gdp_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock gdp_mutex &k->list_lock irq_context: 0 &hba->device_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock uevent_sock_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock running_helpers_waitq.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#37 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#37 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock cgwb_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pin_fs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock init_fs.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bdi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock inode_hash_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bdev_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex free_vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#4 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &mapping->i_private_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &base->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &folio_wait_table[i] irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 part_parser_lock irq_context: 0 mtd_table_mutex irq_context: 0 chip_drvs_lock irq_context: hardirq &ret->b_uptodate_lock irq_context: hardirq &folio_wait_table[i] irq_context: hardirq &folio_wait_table[i] &p->pi_lock irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 pool_lock#2 irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (kmod_concurrent_max).lock irq_context: 0 &x->wait#17 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sig->wait_chldexit irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &prev->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock irq_context: 0 mtd_table_mutex fs_reclaim irq_context: 0 mtd_table_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex pool_lock#2 irq_context: 0 mtd_table_mutex &x->wait#9 irq_context: 0 mtd_table_mutex &obj_hash[i].lock irq_context: 0 mtd_table_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex irq_context: 0 mtd_table_mutex gdp_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex gdp_mutex pool_lock#2 irq_context: 0 mtd_table_mutex gdp_mutex lock irq_context: 0 mtd_table_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex lock irq_context: 0 mtd_table_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex bus_type_sem irq_context: 0 mtd_table_mutex sysfs_symlink_target_lock irq_context: 0 mtd_table_mutex &c->lock irq_context: 0 mtd_table_mutex &____s->seqcount irq_context: 0 mtd_table_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &dev->power.lock irq_context: 0 mtd_table_mutex dpm_list_mtx irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 &____s->seqcount irq_context: 0 mtd_table_mutex req_lock irq_context: 0 mtd_table_mutex &p->pi_lock irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex &x->wait#11 irq_context: 0 mtd_table_mutex &rq->__lock irq_context: 0 mtd_table_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex uevent_sock_mutex irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 mtd_table_mutex running_helpers_waitq.lock irq_context: 0 mtd_table_mutex subsys mutex#48 irq_context: 0 mtd_table_mutex subsys mutex#48 &k->k_lock irq_context: 0 mtd_table_mutex devtree_lock irq_context: 0 mtd_table_mutex &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mtd_table_mutex nvmem_ida.xa_lock irq_context: 0 mtd_table_mutex nvmem_cell_mutex irq_context: 0 mtd_table_mutex &k->k_lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex &dev->mutex &dev->power.lock irq_context: 0 mtd_table_mutex &dev->mutex &k->list_lock irq_context: 0 mtd_table_mutex &dev->mutex &k->k_lock irq_context: 0 mtd_table_mutex subsys mutex#49 irq_context: 0 mtd_table_mutex nvmem_mutex irq_context: 0 mtd_table_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex (console_sem).lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner_lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 mtd_table_mutex pcpu_alloc_mutex irq_context: 0 mtd_table_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex batched_entropy_u32.lock irq_context: 0 mtd_table_mutex mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex blk_queue_ida.xa_lock irq_context: 0 mtd_table_mutex &obj_hash[i].lock pool_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock &xa->xa_lock#11 irq_context: 0 mtd_table_mutex &set->tag_list_lock irq_context: 0 mtd_table_mutex bio_slab_lock irq_context: 0 mtd_table_mutex percpu_counters_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_lock_key#3 irq_context: 0 mtd_table_mutex &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &xa->xa_lock#10 irq_context: 0 mtd_table_mutex lock &q->queue_lock irq_context: 0 mtd_table_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex set->srcu irq_context: 0 mtd_table_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->queue_lock irq_context: 0 mtd_table_mutex &q->queue_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->queue_lock pcpu_lock irq_context: 0 mtd_table_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->queue_lock percpu_counters_lock irq_context: 0 mtd_table_mutex &q->queue_lock &blkcg->lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &bdev->bd_size_lock irq_context: 0 mtd_table_mutex elv_list_lock irq_context: 0 mtd_table_mutex (work_completion)(&(&q->requeue_work)->work) irq_context: 0 mtd_table_mutex (work_completion)(&(&hctx->run_work)->work) irq_context: 0 mtd_table_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex subsys mutex#36 irq_context: 0 mtd_table_mutex subsys mutex#36 &k->k_lock irq_context: 0 mtd_table_mutex dev_hotplug_mutex irq_context: 0 mtd_table_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 mtd_table_mutex subsys mutex#37 irq_context: 0 mtd_table_mutex subsys mutex#37 &k->k_lock irq_context: 0 mtd_table_mutex cgwb_lock irq_context: 0 mtd_table_mutex bdi_lock irq_context: 0 mtd_table_mutex inode_hash_lock irq_context: 0 mtd_table_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 rtnl_mutex stack_depot_init_mutex irq_context: 0 rtnl_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex kthread_create_lock irq_context: 0 rtnl_mutex &p->pi_lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &x->wait irq_context: 0 rtnl_mutex wq_pool_mutex irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex crngs.lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 irq_context: 0 rtnl_mutex net_rwsem irq_context: 0 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &x->wait#9 irq_context: 0 rtnl_mutex &k->list_lock irq_context: 0 rtnl_mutex gdp_mutex irq_context: 0 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex bus_type_sem irq_context: 0 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &dev->power.lock irq_context: 0 rtnl_mutex dpm_list_mtx irq_context: 0 rtnl_mutex uevent_sock_mutex irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex running_helpers_waitq.lock irq_context: 0 rtnl_mutex subsys mutex#20 irq_context: 0 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 rtnl_mutex &dir->lock#2 irq_context: 0 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex (console_sem).lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_hotplug_mutex irq_context: 0 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex dev_base_lock irq_context: 0 rtnl_mutex input_pool.lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock irq_context: 0 rtnl_mutex &tbl->lock irq_context: 0 rtnl_mutex sysctl_lock irq_context: 0 rtnl_mutex nl_table_lock irq_context: 0 rtnl_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: 0 rtnl_mutex lweventlist_lock irq_context: 0 rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)gid-cache-wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 once_lock irq_context: 0 once_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) pool_lock#2 irq_context: 0 (inet6addr_validator_chain).rwsem irq_context: 0 (inetaddr_validator_chain).rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_owner_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_owner irq_context: 0 console_owner_lock irq_context: 0 console_owner irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &bdev->bd_size_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &x->wait#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex ext_devt_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &k->list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex bus_type_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex dpm_list_mtx irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex req_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &x->wait#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex subsys mutex#36 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex subsys mutex#36 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#10 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex inode_hash_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex purge_vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#5 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &lruvec->lru_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock nvme_ns_chr_minor_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock chrdevs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#50 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#50 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#51 irq_context: 0 subsys mutex#51 &k->k_lock irq_context: 0 gpio_lookup_lock irq_context: 0 mdio_board_lock irq_context: 0 mode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 l3mdev_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex fs_reclaim &rq->__lock irq_context: 0 &dev->mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &dev->mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &dev->mutex fs_reclaim pool_lock#2 irq_context: softirq rcu_callback quarantine_lock irq_context: 0 &dev->mutex init_mm.page_table_lock irq_context: 0 &dev->mutex stack_depot_init_mutex irq_context: 0 &dev->mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex cpu_hotplug_lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex wq_pool_mutex irq_context: 0 &dev->mutex wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex pools_reg_lock irq_context: 0 &dev->mutex pools_reg_lock pools_lock irq_context: 0 &dev->mutex pools_reg_lock fs_reclaim irq_context: 0 &dev->mutex pools_reg_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex pools_reg_lock pool_lock#2 irq_context: 0 &dev->mutex pools_reg_lock lock irq_context: 0 &dev->mutex pools_reg_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex pools_reg_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex pools_reg_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &retval->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock quarantine_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex &____s->seqcount irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock tmpmask_lock irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock &its->lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock &its->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rtnl_mutex irq_context: 0 &dev->mutex rtnl_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex &xa->xa_lock#4 irq_context: 0 &dev->mutex rtnl_mutex net_rwsem irq_context: 0 &dev->mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 &dev->mutex rtnl_mutex &x->wait#9 irq_context: 0 &dev->mutex rtnl_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 &dev->mutex rtnl_mutex lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex bus_type_sem irq_context: 0 &dev->mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rtnl_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dpm_list_mtx irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rtnl_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rtnl_mutex &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#20 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex &dir->lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dev_base_lock irq_context: 0 &dev->mutex rtnl_mutex input_pool.lock irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 &dev->mutex rtnl_mutex &tbl->lock irq_context: 0 &dev->mutex rtnl_mutex sysctl_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)gve irq_context: 0 (wq_completion)gve &rq->__lock irq_context: 0 (wq_completion)gve &cfs_rq->removed.lock irq_context: 0 (wq_completion)gve &obj_hash[i].lock irq_context: 0 (wq_completion)gve pool_lock#2 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &rq->__lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &cfs_rq->removed.lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &obj_hash[i].lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) pool_lock#2 irq_context: 0 hnae3_common_lock irq_context: 0 &sb->s_type->i_mutex_key &rq->__lock irq_context: 0 subsys mutex#52 irq_context: 0 subsys mutex#52 &k->k_lock irq_context: 0 compressor_list_lock irq_context: 0 compressor_list_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 hwsim_radio_lock irq_context: 0 subsys mutex#53 irq_context: 0 subsys mutex#53 &k->k_lock irq_context: 0 deferred_probe_mutex irq_context: 0 rtnl_mutex param_lock irq_context: 0 rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 rtnl_mutex (console_sem).lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx running_helpers_waitq.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 rtnl_mutex &base->lock irq_context: 0 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex irq_context: 0 rfkill_global_mutex fs_reclaim irq_context: 0 rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rfkill_global_mutex pool_lock#2 irq_context: 0 rfkill_global_mutex &k->list_lock irq_context: 0 rfkill_global_mutex lock irq_context: 0 rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rfkill_global_mutex bus_type_sem irq_context: 0 rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 rfkill_global_mutex &c->lock irq_context: 0 rfkill_global_mutex &____s->seqcount irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex &dev->power.lock irq_context: 0 rfkill_global_mutex dpm_list_mtx irq_context: 0 rfkill_global_mutex &rfkill->lock irq_context: 0 rfkill_global_mutex uevent_sock_mutex irq_context: 0 rfkill_global_mutex &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 rfkill_global_mutex &k->k_lock irq_context: 0 rfkill_global_mutex subsys mutex#40 irq_context: 0 rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 rfkill_global_mutex triggers_list_lock irq_context: 0 rfkill_global_mutex leds_list_lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex &rq->__lock irq_context: 0 rfkill_global_mutex.wait_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#4 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex crngs.lock irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 &dev->mutex rtnl_mutex crngs.lock irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx &sec->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex rtnl_mutex &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex &local->iflist_mtx#2 irq_context: 0 &dev->mutex hwsim_phys_lock irq_context: 0 &dev->mutex nl_table_lock irq_context: 0 &dev->mutex nl_table_wait.lock irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex hwsim_phys_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 xdomain_lock irq_context: 0 xdomain_lock fs_reclaim irq_context: 0 xdomain_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 xdomain_lock pool_lock#2 irq_context: 0 ioctl_mutex irq_context: 0 address_handler_list_lock irq_context: 0 card_mutex irq_context: 0 subsys mutex#56 irq_context: 0 subsys mutex#56 &k->k_lock irq_context: 0 &x->wait#18 irq_context: 0 &x->wait#18 &p->pi_lock irq_context: 0 &x->wait#18 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &txlock irq_context: 0 &txlock &list->lock#3 irq_context: 0 &txlock &txwq irq_context: 0 &iocq[i].lock irq_context: 0 &iocq[i].lock &ktiowq[i] irq_context: 0 &txwq irq_context: 0 &txwq &p->pi_lock irq_context: 0 &txwq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh pool_lock#2 irq_context: 0 subsys mutex#57 irq_context: 0 subsys mutex#57 &k->k_lock irq_context: 0 usb_bus_idr_lock irq_context: 0 usb_bus_idr_lock (usb_notifier_list).rwsem irq_context: 0 table_lock irq_context: 0 table_lock &k->list_lock irq_context: 0 table_lock fs_reclaim irq_context: 0 table_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 table_lock pool_lock#2 irq_context: 0 table_lock lock irq_context: 0 table_lock lock kernfs_idr_lock irq_context: 0 table_lock &root->kernfs_rwsem irq_context: 0 table_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 table_lock &k->k_lock irq_context: 0 table_lock uevent_sock_mutex irq_context: 0 table_lock &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock irq_context: 0 table_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 table_lock running_helpers_waitq.lock irq_context: 0 table_lock (console_sem).lock irq_context: 0 table_lock console_lock console_srcu console_owner_lock irq_context: 0 table_lock console_lock console_srcu console_owner irq_context: 0 table_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 table_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 table_lock &c->lock irq_context: 0 table_lock &pcp->lock &zone->lock irq_context: 0 table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 table_lock &____s->seqcount irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 table_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 table_lock &rq->__lock irq_context: 0 table_lock &obj_hash[i].lock pool_lock irq_context: softirq (&ipmi_timer) irq_context: softirq (&ipmi_timer) &ipmi_interfaces_srcu irq_context: 0 table_lock batched_entropy_u8.lock irq_context: 0 table_lock kfence_freelist_lock irq_context: 0 &dev->mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pin_fs_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#9 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem bus_type_sem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &dev->power.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem dpm_list_mtx irq_context: 0 &dev->mutex (usb_notifier_list).rwsem req_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#11 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem uevent_sock_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#57 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#57 &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem mon_lock irq_context: 0 &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock irq_context: softirq &bh->lock irq_context: softirq lock#6 irq_context: softirq lock#6 kcov_remote_lock irq_context: softirq &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock input_pool.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock req_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#11 irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex set_config_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &new_driver->dynids.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pm_qos_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex component_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex subsys mutex#58 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock subsys mutex#58 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#9 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &hub->irq_urb_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&hub->irq_urb_retry) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_unlink_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) usb_kill_urb_queue.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq usb_kill_urb_queue.lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock irq_context: 0 &dev->mutex udc_lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex gdp_mutex &c->lock irq_context: 0 &dev->mutex gdp_mutex &____s->seqcount irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (work_completion)(&hub->tt.clear_work) irq_context: 0 &dev->mutex subsys mutex#59 irq_context: 0 &dev->mutex subsys mutex#59 &k->k_lock irq_context: 0 &dev->mutex gadget_id_numbers.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events kernfs_notify_work irq_context: 0 (wq_completion)events kernfs_notify_work kernfs_notify_lock irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem irq_context: 0 &dev->mutex subsys mutex#60 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dum_hcd->dum->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) device_state_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_list_lock irq_context: 0 func_lock irq_context: 0 g_tf_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &lock->wait_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &x->wait#19 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&timer.timer) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &meta->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex quarantine_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock pool_lock irq_context: 0 reading_mutex &rq->__lock irq_context: 0 reading_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &x->wait#14 &p->pi_lock irq_context: hardirq &x->wait#14 &p->pi_lock &rq->__lock irq_context: hardirq &x->wait#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &meta->lock irq_context: hardirq &x->wait#4 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &cfs_rq->removed.lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock quarantine_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex usb_bus_idr_lock kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex &dev->mutex &lock->wait_lock irq_context: 0 &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 input_ida.xa_lock irq_context: 0 input_ida.xa_lock pool_lock#2 irq_context: 0 subsys mutex#31 irq_context: 0 subsys mutex#31 &k->k_lock irq_context: 0 input_mutex input_ida.xa_lock irq_context: 0 input_mutex fs_reclaim irq_context: 0 input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 input_mutex pool_lock#2 irq_context: 0 input_mutex &x->wait#9 irq_context: 0 input_mutex &obj_hash[i].lock irq_context: 0 input_mutex &dev->mutex#2 irq_context: 0 input_mutex chrdevs_lock irq_context: 0 input_mutex &k->list_lock irq_context: 0 input_mutex lock irq_context: 0 input_mutex lock kernfs_idr_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 input_mutex bus_type_sem irq_context: 0 input_mutex sysfs_symlink_target_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &dev->power.lock irq_context: 0 input_mutex dpm_list_mtx irq_context: 0 input_mutex req_lock irq_context: 0 input_mutex &p->pi_lock irq_context: 0 input_mutex &x->wait#11 irq_context: 0 input_mutex &rq->__lock irq_context: 0 input_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex uevent_sock_mutex irq_context: 0 input_mutex rcu_read_lock &pool->lock irq_context: 0 input_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 input_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 input_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 input_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 input_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex running_helpers_waitq.lock irq_context: 0 input_mutex &k->k_lock irq_context: 0 input_mutex subsys mutex#31 irq_context: 0 input_mutex subsys mutex#31 &k->k_lock irq_context: 0 input_mutex &c->lock irq_context: 0 input_mutex &____s->seqcount irq_context: 0 serio_event_lock irq_context: 0 serio_event_lock pool_lock#2 irq_context: 0 serio_event_lock rcu_read_lock &pool->lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 serio_event_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long irq_context: 0 (wq_completion)events_long serio_event_work irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex serio_event_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex pool_lock#2 irq_context: 0 &new_driver->dynids.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (efi_runtime_lock).lock irq_context: 0 &dev->mutex &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtc_ida.xa_lock irq_context: 0 &dev->mutex &rtc->ops_lock irq_context: 0 &dev->mutex &rtc->ops_lock (efi_runtime_lock).lock irq_context: 0 &dev->mutex &rtc->ops_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &rtc->ops_lock &x->wait#12 irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &rq->__lock irq_context: 0 &dev->mutex chrdevs_lock irq_context: 0 &dev->mutex req_lock irq_context: 0 &dev->mutex &x->wait#11 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#27 irq_context: 0 &dev->mutex subsys mutex#27 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &x->wait#9 irq_context: 0 &dev->mutex subsys mutex#27 &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex subsys mutex#27 platform_devid_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 lock irq_context: 0 &dev->mutex subsys mutex#27 lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 bus_type_sem irq_context: 0 &dev->mutex subsys mutex#27 sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &c->lock irq_context: 0 &dev->mutex subsys mutex#27 &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 dpm_list_mtx irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 &dev->mutex subsys mutex#27 &pcp->lock &zone->lock irq_context: 0 &dev->mutex subsys mutex#27 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#4 irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#27 wakeup_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#13 irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 events_lock irq_context: 0 &dev->mutex subsys mutex#27 rtcdev_lock irq_context: 0 g_smscore_deviceslock irq_context: 0 g_smscore_deviceslock fs_reclaim irq_context: 0 g_smscore_deviceslock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 g_smscore_deviceslock pool_lock#2 irq_context: 0 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim &rq->__lock irq_context: 0 fs_reclaim &cfs_rq->removed.lock irq_context: 0 fs_reclaim &obj_hash[i].lock irq_context: 0 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cx231xx_devlist_mutex irq_context: 0 em28xx_devlist_mutex irq_context: 0 pvr2_context_sync_data.lock irq_context: 0 &dev->mutex core_lock irq_context: 0 &dev->mutex core_lock fs_reclaim irq_context: 0 &dev->mutex core_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex core_lock pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem i2c_dev_list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#9 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem chrdevs_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem bus_type_sem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &dev->power.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem dpm_list_mtx irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem req_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#11 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem uevent_sock_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->k_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#61 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#61 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#62 irq_context: 0 &dev->mutex pin_fs_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex core_lock &k->list_lock irq_context: 0 &dev->mutex core_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock irq_context: 0 &dev->mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex (kmod_concurrent_max).lock irq_context: 0 &dev->mutex &x->wait#17 irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &dev->mutex &c->lock irq_context: 0 &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex frontend_mutex irq_context: 0 &dev->mutex frontend_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &xa->xa_lock#13 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#63 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#63 &k->k_lock irq_context: 0 &dev->mutex &dmxdev->lock irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#13 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#13 &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#13 &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#13 pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#63 irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#63 &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &dvbdemux->mutex irq_context: 0 &dev->mutex media_devnode_lock irq_context: 0 &dev->mutex subsys mutex#64 irq_context: 0 &dev->mutex videodev_lock irq_context: 0 &dev->mutex subsys mutex#65 irq_context: 0 &dev->mutex subsys mutex#65 &k->k_lock irq_context: 0 &dev->mutex &xa->xa_lock#13 irq_context: 0 &dev->mutex &mdev->graph_mutex irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex vimc_sensor:396:(&vsensor->hdl)->_lock irq_context: 0 &dev->mutex &v4l2_dev->lock irq_context: 0 &dev->mutex vimc_debayer:581:(&vdebayer->hdl)->_lock irq_context: 0 &dev->mutex vimc_lens:61:(&vlens->hdl)->_lock irq_context: 0 &dev->mutex tk_core.seq.seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1618:(hdl_fb)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &obj_hash[i].lock irq_context: 0 &adap->kthread_waitq irq_context: 0 &dev->cec_xfers_slock irq_context: 0 &dev->kthread_waitq_cec irq_context: 0 &dev->mutex cec_devnode_lock irq_context: 0 &dev->mutex subsys mutex#66 irq_context: 0 &dev->mutex &adap->lock irq_context: 0 &dev->mutex &adap->lock tk_core.seq.seqcount irq_context: 0 &dev->mutex &adap->lock &adap->devnode.lock_fhs irq_context: 0 &dev->mutex gdp_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex gdp_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 ptp_clocks_map.xa_lock irq_context: 0 subsys mutex#67 irq_context: 0 subsys mutex#67 &k->k_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pers_lock irq_context: 0 _lock irq_context: 0 dm_bufio_clients_lock irq_context: 0 _ps_lock irq_context: 0 _lock#2 irq_context: 0 _lock#3 irq_context: 0 register_lock#2 irq_context: 0 subsys mutex#68 irq_context: 0 subsys mutex#68 &k->k_lock irq_context: 0 bp_lock irq_context: 0 bp_lock irq_context: 0 subsys mutex#69 irq_context: 0 subsys mutex#69 &k->k_lock irq_context: 0 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 free_vmap_area_lock pool_lock#2 irq_context: softirq (&dsp_spl_tl) irq_context: softirq (&dsp_spl_tl) dsp_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock tk_core.seq.seqcount irq_context: softirq (&dsp_spl_tl) dsp_lock &obj_hash[i].lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock &obj_hash[i].lock irq_context: 0 &x->wait#6 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex lock#7 irq_context: 0 iscsi_transport_lock irq_context: 0 subsys mutex#70 irq_context: 0 subsys mutex#70 &k->k_lock irq_context: 0 &tx_task->waiting irq_context: 0 link_ops_rwsem irq_context: 0 disable_lock irq_context: 0 disable_lock fs_reclaim irq_context: 0 disable_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 disable_lock pool_lock#2 irq_context: 0 disable_lock &x->wait#9 irq_context: 0 disable_lock &obj_hash[i].lock irq_context: 0 disable_lock &k->list_lock irq_context: 0 disable_lock &c->lock irq_context: 0 disable_lock &____s->seqcount irq_context: 0 disable_lock lock irq_context: 0 disable_lock lock kernfs_idr_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 disable_lock bus_type_sem irq_context: 0 disable_lock sysfs_symlink_target_lock irq_context: 0 disable_lock &k->k_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &dev->power.lock irq_context: 0 disable_lock dpm_list_mtx irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 disable_lock uevent_sock_mutex irq_context: 0 disable_lock rcu_read_lock &pool->lock irq_context: 0 disable_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 disable_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 disable_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 disable_lock running_helpers_waitq.lock irq_context: 0 disable_lock &dev->mutex &dev->power.lock irq_context: 0 disable_lock &dev->mutex &k->list_lock irq_context: 0 disable_lock &dev->mutex &k->k_lock irq_context: 0 disable_lock subsys mutex#4 irq_context: 0 protocol_lock irq_context: 0 protocol_lock pool_lock#2 irq_context: 0 psinfo_lock irq_context: 0 psinfo_lock fs_reclaim irq_context: 0 psinfo_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 psinfo_lock pool_lock#2 irq_context: 0 psinfo_lock free_vmap_area_lock irq_context: 0 psinfo_lock vmap_area_lock irq_context: 0 psinfo_lock &____s->seqcount irq_context: 0 psinfo_lock init_mm.page_table_lock irq_context: 0 psinfo_lock (console_sem).lock irq_context: 0 psinfo_lock console_lock console_srcu console_owner_lock irq_context: 0 psinfo_lock console_lock console_srcu console_owner irq_context: 0 psinfo_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 psinfo_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 psinfo_lock pstore_sb_lock irq_context: 0 psinfo_lock dump_list_lock irq_context: 0 psinfo_lock &rq->__lock irq_context: 0 psinfo_lock &cfs_rq->removed.lock irq_context: 0 psinfo_lock &obj_hash[i].lock irq_context: 0 vsock_register_mutex irq_context: 0 comedi_drivers_list_lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &domain->mutex irq_context: 0 &domain->mutex sparse_irq_lock irq_context: 0 &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 &domain->mutex sparse_irq_lock lock irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &domain->mutex fs_reclaim irq_context: 0 &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &domain->mutex pool_lock#2 irq_context: 0 &domain->mutex &irq_desc_lock_class irq_context: 0 &desc->request_mutex irq_context: 0 &desc->request_mutex &irq_desc_lock_class irq_context: 0 &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 pmus_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cscfg_mutex irq_context: 0 cscfg_mutex fs_reclaim irq_context: 0 cscfg_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cscfg_mutex &c->lock irq_context: 0 cscfg_mutex &____s->seqcount irq_context: 0 cscfg_mutex pool_lock#2 irq_context: 0 cscfg_mutex &x->wait#9 irq_context: 0 cscfg_mutex &obj_hash[i].lock irq_context: 0 cscfg_mutex &k->list_lock irq_context: 0 cscfg_mutex lock irq_context: 0 cscfg_mutex lock kernfs_idr_lock irq_context: 0 cscfg_mutex &root->kernfs_rwsem irq_context: 0 cscfg_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cscfg_mutex bus_type_sem irq_context: 0 cscfg_mutex &root->kernfs_rwsem irq_context: 0 cscfg_mutex &dev->power.lock irq_context: 0 cscfg_mutex dpm_list_mtx irq_context: 0 fs_reclaim icc_bw_lock irq_context: 0 subsys mutex#71 irq_context: 0 subsys mutex#71 &k->k_lock irq_context: 0 snd_ctl_layer_rwsem irq_context: 0 snd_card_mutex irq_context: 0 snd_ioctl_rwsem irq_context: 0 strings irq_context: 0 strings fs_reclaim irq_context: 0 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 strings pool_lock#2 irq_context: 0 register_mutex irq_context: 0 sound_mutex irq_context: 0 sound_mutex fs_reclaim irq_context: 0 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sound_mutex pool_lock#2 irq_context: 0 sound_mutex &k->list_lock irq_context: 0 sound_mutex gdp_mutex irq_context: 0 sound_mutex gdp_mutex &k->list_lock irq_context: 0 sound_mutex lock irq_context: 0 sound_mutex lock kernfs_idr_lock irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sound_mutex bus_type_sem irq_context: 0 sound_mutex sysfs_symlink_target_lock irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex &c->lock irq_context: 0 sound_mutex &____s->seqcount irq_context: 0 sound_mutex &dev->power.lock irq_context: 0 sound_mutex dpm_list_mtx irq_context: 0 sound_mutex req_lock irq_context: 0 sound_mutex &p->pi_lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sound_mutex &rq->__lock irq_context: 0 sound_mutex &x->wait#11 irq_context: 0 sound_mutex &obj_hash[i].lock irq_context: 0 sound_mutex uevent_sock_mutex irq_context: 0 sound_mutex rcu_read_lock &pool->lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sound_mutex running_helpers_waitq.lock irq_context: 0 sound_mutex subsys mutex#71 irq_context: 0 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 register_mutex#2 irq_context: 0 register_mutex#3 irq_context: 0 register_mutex#3 fs_reclaim irq_context: 0 register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex irq_context: 0 register_mutex#3 sound_mutex fs_reclaim irq_context: 0 register_mutex#3 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 sound_mutex pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex gdp_mutex irq_context: 0 register_mutex#3 sound_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex lock irq_context: 0 register_mutex#3 sound_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#3 sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#3 sound_mutex bus_type_sem irq_context: 0 register_mutex#3 sound_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &dev->power.lock irq_context: 0 register_mutex#3 sound_mutex dpm_list_mtx irq_context: 0 register_mutex#3 sound_mutex req_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 sound_mutex &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &x->wait#11 irq_context: 0 register_mutex#3 sound_mutex &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex uevent_sock_mutex irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#3 sound_mutex subsys mutex#71 irq_context: 0 register_mutex#3 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 register_mutex#3 clients_lock irq_context: 0 &client->ports_mutex irq_context: 0 &client->ports_mutex &client->ports_lock irq_context: 0 register_mutex#4 irq_context: 0 register_mutex#4 fs_reclaim irq_context: 0 register_mutex#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 sound_oss_mutex pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex sound_loader_lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#9 irq_context: 0 register_mutex#4 sound_oss_mutex &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex &c->lock irq_context: 0 register_mutex#4 sound_oss_mutex &____s->seqcount irq_context: 0 register_mutex#4 sound_oss_mutex lock irq_context: 0 register_mutex#4 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex bus_type_sem irq_context: 0 register_mutex#4 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &dev->power.lock irq_context: 0 register_mutex#4 sound_oss_mutex dpm_list_mtx irq_context: 0 register_mutex#4 sound_oss_mutex req_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#11 irq_context: 0 register_mutex#4 sound_oss_mutex uevent_sock_mutex irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#71 irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 clients_lock irq_context: 0 &client->ports_lock irq_context: 0 &grp->list_mutex/1 irq_context: 0 &grp->list_mutex#2 irq_context: 0 &grp->list_mutex#2 &grp->list_lock irq_context: 0 &grp->list_mutex/1 clients_lock irq_context: 0 &grp->list_mutex/1 &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events async_lookup_work pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work clients_lock irq_context: 0 (wq_completion)events async_lookup_work &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work snd_card_mutex irq_context: 0 (wq_completion)events async_lookup_work (kmod_concurrent_max).lock irq_context: 0 (wq_completion)events async_lookup_work &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events async_lookup_work &x->wait#17 irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 &c->lock irq_context: 0 register_mutex#3 &____s->seqcount irq_context: 0 &grp->list_mutex/1 register_lock#3 irq_context: 0 &grp->list_mutex/1 fs_reclaim irq_context: 0 &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex snd_card_mutex irq_context: 0 &dev->mutex &entry->access irq_context: 0 &dev->mutex info_mutex irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 &dev->mutex info_mutex fs_reclaim irq_context: 0 &dev->mutex info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex info_mutex &____s->seqcount irq_context: 0 &dev->mutex info_mutex pool_lock#2 irq_context: 0 &dev->mutex info_mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 &pcp->lock &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#14 &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem &card->ctl_files_rwlock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex subsys mutex#71 irq_context: 0 &dev->mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work running_helpers_waitq.lock irq_context: 0 (wq_completion)events async_lookup_work autoload_work irq_context: 0 (wq_completion)events async_lookup_work &x->wait#10 irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events autoload_work irq_context: 0 (wq_completion)events autoload_work &k->list_lock irq_context: 0 (wq_completion)events autoload_work &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#71 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 &c->lock irq_context: 0 &dev->mutex register_mutex#2 &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 register_mutex irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#71 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 strings irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 strings pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_mutex#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 &entry->access irq_context: 0 &dev->mutex register_mutex#2 info_mutex irq_context: 0 &dev->mutex sound_mutex irq_context: 0 &dev->mutex sound_mutex fs_reclaim irq_context: 0 &dev->mutex sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_mutex &k->list_lock irq_context: 0 &dev->mutex sound_mutex lock irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_mutex bus_type_sem irq_context: 0 &dev->mutex sound_mutex &c->lock irq_context: 0 &dev->mutex sound_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_mutex req_lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_mutex &rq->__lock irq_context: 0 &dev->mutex sound_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_mutex &k->k_lock irq_context: 0 &dev->mutex sound_mutex subsys mutex#71 irq_context: 0 &dev->mutex sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem snd_ctl_led_mutex irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &x->wait#9 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->list_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem bus_type_sem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &dev->power.lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem dpm_list_mtx irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex info_mutex &c->lock irq_context: 0 &dev->mutex info_mutex &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex sound_oss_mutex lock irq_context: 0 &dev->mutex sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_oss_mutex &c->lock irq_context: 0 &dev->mutex sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_oss_mutex req_lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#71 irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex strings irq_context: 0 &dev->mutex strings fs_reclaim irq_context: 0 &dev->mutex strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex strings pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex strings &____s->seqcount irq_context: 0 &dev->mutex register_mutex#5 irq_context: 0 &dev->mutex register_mutex#3 irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#3 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#3 clients_lock irq_context: 0 &dev->mutex clients_lock irq_context: 0 &dev->mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 irq_context: 0 &dev->mutex &grp->list_mutex/1 clients_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &client->ports_lock irq_context: 0 &dev->mutex &client->ports_mutex irq_context: 0 &dev->mutex &client->ports_mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 register_lock#3 irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex sound_oss_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex strings &c->lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &c->lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &____s->seqcount irq_context: 0 &dev->mutex sound_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex sound_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex client_mutex irq_context: 0 &dev->mutex client_mutex fs_reclaim irq_context: 0 &dev->mutex client_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex client_mutex pool_lock#2 irq_context: 0 &dev->mutex client_mutex &dev->devres_lock irq_context: 0 &dev->mutex client_mutex &c->lock irq_context: 0 &dev->mutex client_mutex &____s->seqcount irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 llc_sap_list_lock irq_context: 0 llc_sap_list_lock pool_lock#2 irq_context: 0 act_id_mutex irq_context: 0 act_id_mutex fs_reclaim irq_context: 0 act_id_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 act_id_mutex pool_lock#2 irq_context: 0 act_mod_lock irq_context: 0 act_id_mutex &____s->seqcount irq_context: 0 ife_mod_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cls_mod_lock irq_context: 0 ematch_mod_lock irq_context: 0 sock_diag_table_mutex irq_context: 0 nfnl_subsys_acct irq_context: 0 nfnl_subsys_queue irq_context: 0 nfnl_subsys_ulog irq_context: 0 nf_log_mutex irq_context: 0 nfnl_subsys_osf irq_context: 0 nf_sockopt_mutex irq_context: 0 nfnl_subsys_ctnetlink irq_context: 0 nfnl_subsys_ctnetlink_exp irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 nfnl_subsys_cttimeout irq_context: 0 nfnl_subsys_cthelper irq_context: 0 nf_ct_helper_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 nf_conntrack_expect_lock irq_context: 0 nf_ct_nat_helpers_mutex irq_context: 0 nfnl_subsys_nftables irq_context: 0 nfnl_subsys_nftcompat irq_context: 0 masq_mutex irq_context: 0 masq_mutex pernet_ops_rwsem irq_context: 0 masq_mutex pernet_ops_rwsem rtnl_mutex irq_context: 0 masq_mutex (inetaddr_chain).rwsem irq_context: 0 masq_mutex inet6addr_chain.lock irq_context: 0 &xt[i].mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 subsys mutex#72 irq_context: 0 subsys mutex#72 &k->k_lock irq_context: 0 nfnl_subsys_ipset irq_context: 0 ip_set_type_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 ip_vs_sched_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 ip_vs_pe_mutex irq_context: 0 tunnel4_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 xfrm4_protocol_mutex irq_context: 0 &xt[i].mutex fs_reclaim irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &xt[i].mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock pool_lock#2 irq_context: 0 &xt[i].mutex &c->lock irq_context: 0 &xt[i].mutex &____s->seqcount irq_context: 0 inet_diag_table_mutex irq_context: 0 xfrm_km_lock irq_context: 0 xfrm6_protocol_mutex irq_context: 0 tunnel6_mutex irq_context: 0 xfrm_if_cb_lock irq_context: 0 inetsw6_lock irq_context: 0 &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 (crypto_chain).rwsem fs_reclaim irq_context: 0 (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (crypto_chain).rwsem pool_lock#2 irq_context: 0 (crypto_chain).rwsem kthread_create_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock irq_context: 0 (crypto_chain).rwsem &x->wait irq_context: 0 (crypto_chain).rwsem &rq->__lock irq_context: 0 (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 &x->wait#20 irq_context: 0 &x->wait#20 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait irq_context: 0 (crypto_chain).rwsem &c->lock irq_context: 0 (crypto_chain).rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 stp_proto_mutex irq_context: 0 stp_proto_mutex llc_sap_list_lock irq_context: 0 stp_proto_mutex llc_sap_list_lock &c->lock irq_context: 0 stp_proto_mutex llc_sap_list_lock &____s->seqcount irq_context: 0 stp_proto_mutex llc_sap_list_lock pool_lock#2 irq_context: 0 switchdev_notif_chain.lock irq_context: 0 (switchdev_blocking_notif_chain).rwsem irq_context: 0 br_ioctl_mutex irq_context: 0 nf_ct_proto_mutex irq_context: 0 ebt_mutex irq_context: 0 ebt_mutex fs_reclaim irq_context: 0 ebt_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex pool_lock#2 irq_context: 0 dsa_tag_drivers_lock irq_context: 0 rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 protocol_list_lock irq_context: 0 linkfail_lock irq_context: 0 fs_reclaim &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex quarantine_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rose_neigh_list_lock irq_context: 0 proto_tab_lock#2 irq_context: 0 lock pidmap_lock &c->lock irq_context: 0 lock pidmap_lock &____s->seqcount irq_context: 0 bt_proto_lock irq_context: 0 bt_proto_lock pool_lock#2 irq_context: 0 bt_proto_lock &dir->lock irq_context: 0 bt_proto_lock &obj_hash[i].lock irq_context: 0 bt_proto_lock chan_list_lock irq_context: 0 bt_proto_lock l2cap_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP chan_list_lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_wq.lock irq_context: 0 rfcomm_mutex irq_context: 0 auth_domain_lock irq_context: 0 registered_mechs_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 atm_dev_notify_chain.lock irq_context: 0 genl_mutex irq_context: 0 proto_tab_lock#3 irq_context: 0 vlan_ioctl_mutex irq_context: 0 pernet_ops_rwsem (console_sem).lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner_lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rds_info_lock irq_context: 0 &wq->mutex &pool->lock &p->pi_lock irq_context: 0 rds_trans_sem irq_context: 0 rds_trans_sem (console_sem).lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner_lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner irq_context: 0 rds_trans_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 rds_trans_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 &id_priv->lock irq_context: 0 lock#7 irq_context: 0 lock#7 fs_reclaim irq_context: 0 lock#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 lock#7 pool_lock#2 irq_context: 0 lock#7 &xa->xa_lock#15 irq_context: 0 lock#7 &xa->xa_lock#15 pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 smc_wr_rx_hash_lock irq_context: 0 v9fs_trans_lock irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 &x->wait#17 &p->pi_lock irq_context: 0 &x->wait#17 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 lowpan_nhc_lock irq_context: 0 rcu_read_lock quarantine_lock irq_context: 0 ovs_mutex irq_context: 0 pernet_ops_rwsem once_lock irq_context: 0 pernet_ops_rwsem once_lock crngs.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock irq_context: 0 kernfs_idr_lock pool_lock#2 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock purge_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock pool_lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) quarantine_lock irq_context: 0 &root->kernfs_rwsem quarantine_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem kthread_create_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &x->wait irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &x->wait#20 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock &obj_hash[i].lock irq_context: 0 &x->wait#20 &p->pi_lock &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem asymmetric_key_parsers_sem (&timer.timer) irq_context: 0 key_types_sem asymmetric_key_parsers_sem &c->lock irq_context: 0 key_types_sem &type->lock_class irq_context: 0 key_types_sem &type->lock_class fs_reclaim irq_context: 0 key_types_sem &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem &type->lock_class pool_lock#2 irq_context: 0 key_types_sem &type->lock_class key_user_lock irq_context: 0 key_types_sem &type->lock_class crngs.lock irq_context: 0 key_types_sem &type->lock_class key_serial_lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex pool_lock#2 irq_context: 0 key_types_sem &type->lock_class ima_keys_lock irq_context: 0 key_types_sem &obj_hash[i].lock irq_context: 0 key_types_sem pool_lock#2 irq_context: 0 slab_mutex lock irq_context: 0 slab_mutex lock kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem irq_context: 0 slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 slab_mutex &k->list_lock irq_context: 0 slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 slab_mutex lock kernfs_idr_lock &c->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slab_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 slab_mutex &obj_hash[i].lock irq_context: 0 slab_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 slab_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 slab_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex crypto_alg_sem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock free_vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock init_mm.page_table_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->page_table_lock irq_context: 0 ptlock_ptr(ptdesc)#2 irq_context: 0 rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 (wq_completion)events netstamp_work irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#21 irq_context: 0 &x->wait#21 irq_context: 0 &x->wait#21 &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem irq_context: 0 bio_slab_lock slab_mutex &k->list_lock irq_context: 0 bio_slab_lock slab_mutex lock irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 init_user_ns.keyring_sem irq_context: 0 init_user_ns.keyring_sem key_user_lock irq_context: 0 init_user_ns.keyring_sem root_key_user.lock irq_context: 0 init_user_ns.keyring_sem fs_reclaim irq_context: 0 init_user_ns.keyring_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 init_user_ns.keyring_sem pool_lock#2 irq_context: 0 init_user_ns.keyring_sem crngs.lock irq_context: 0 init_user_ns.keyring_sem key_serial_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex irq_context: 0 init_user_ns.keyring_sem &type->lock_class irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock root_key_user.lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex pool_lock#2 irq_context: 0 init_user_ns.keyring_sem keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#2 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 template_list irq_context: 0 idr_lock irq_context: 0 key_types_sem &type->lock_class &c->lock irq_context: 0 key_types_sem &type->lock_class &____s->seqcount irq_context: 0 ima_extend_list_mutex irq_context: 0 ima_extend_list_mutex fs_reclaim irq_context: 0 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ima_extend_list_mutex pool_lock#2 irq_context: softirq (&rxnet->peer_keepalive_timer) irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)krxrpcd irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &rxnet->peer_hash_lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &obj_hash[i].lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock &obj_hash[i].lock irq_context: 0 pci_bus_sem irq_context: 0 clk_debug_lock pin_fs_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound deferred_probe_work irq_context: 0 deferred_probe_work irq_context: 0 (wq_completion)events_unbound deferred_probe_work deferred_probe_mutex irq_context: 0 &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 console_mutex &root->kernfs_rwsem irq_context: 0 console_mutex kernfs_notify_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 console_mutex &rq->__lock irq_context: 0 console_mutex (console_sem).lock irq_context: 0 console_mutex console_lock console_srcu console_owner_lock irq_context: 0 console_mutex console_lock console_srcu console_owner irq_context: 0 console_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 console_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 k-sk_lock-AF_INET irq_context: 0 k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-slock-AF_INET#2 irq_context: 0 k-sk_lock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 k-slock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 power_off_handler_list.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem quarantine_lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &obj_hash[i].lock pool_lock irq_context: 0 reg_requests_lock irq_context: 0 (wq_completion)events reg_work irq_context: 0 (wq_completion)events reg_work rtnl_mutex irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events reg_work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) async_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &____s->seqcount#6 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 detector_work irq_context: 0 &wq->mutex &x->wait#10 irq_context: 0 rcu_read_lock &pool->lock (worker)->lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_gpio_deferred_req_irqs_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &rq->__lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 gpd_list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#9 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem bus_type_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem running_helpers_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#22 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#2 irq_context: 0 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 tomoyo_ss &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 tomoyo_ss tomoyo_log_lock irq_context: 0 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 tomoyo_ss &c->lock irq_context: 0 tomoyo_ss &____s->seqcount irq_context: 0 cdev_lock irq_context: 0 tty_mutex (console_sem).lock irq_context: 0 tty_mutex console_lock irq_context: 0 tty_mutex fs_reclaim irq_context: 0 tty_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex pool_lock#2 irq_context: 0 tty_mutex tty_ldiscs_lock irq_context: 0 tty_mutex &obj_hash[i].lock irq_context: 0 tty_mutex &k->list_lock irq_context: 0 tty_mutex &k->k_lock irq_context: 0 tty_mutex &tty->legacy_mutex irq_context: 0 tty_mutex &tty->legacy_mutex &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem init_mm.page_table_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->legacy_mutex irq_context: 0 &tty->legacy_mutex &tty->files_lock irq_context: 0 &tty->legacy_mutex &port->lock irq_context: 0 &tty->legacy_mutex &port->mutex irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex &____s->seqcount irq_context: 0 &tty->legacy_mutex &port->mutex &port_lock_key irq_context: 0 &tty->legacy_mutex &port->mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex register_lock pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_inum_ida.xa_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port_lock_key irq_context: 0 sb_writers#2 irq_context: 0 sb_writers#2 mount_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss file_systems_lock irq_context: 0 tomoyo_ss fs_reclaim irq_context: 0 tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 tomoyo_ss &rq->__lock irq_context: 0 &type->s_umount_key#24/1 irq_context: 0 &type->s_umount_key#24/1 fs_reclaim irq_context: 0 &type->s_umount_key#24/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 pool_lock#2 irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#24/1 shrinker_mutex irq_context: 0 &type->s_umount_key#24/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#24/1 sb_lock irq_context: 0 &type->s_umount_key#24/1 inode_hash_lock irq_context: 0 &type->s_umount_key#24/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#24/1 bdev_lock irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#24/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#24/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#24/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#24/1 &____s->seqcount irq_context: 0 &type->s_umount_key#24/1 &wq->mutex irq_context: 0 &type->s_umount_key#24/1 &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#24/1 &c->lock irq_context: 0 &type->s_umount_key#24/1 kthread_create_lock irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 &x->wait irq_context: 0 &type->s_umount_key#24/1 &rq->__lock irq_context: 0 &type->s_umount_key#24/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#24/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &type->s_umount_key#24/1 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#24/1 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#24/1 lock#4 irq_context: 0 &type->s_umount_key#24/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#24/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#24/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#24/1 &rq->__lock &cfs_rq->removed.lock irq_context: hardirq bit_wait_table + i irq_context: hardirq bit_wait_table + i &p->pi_lock irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->s_umount_key#24/1 &wq->mutex &x->wait#10 irq_context: 0 &type->s_umount_key#24/1 wq_mayday_lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock (worker)->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 &sbi->old_work_lock irq_context: 0 &type->s_umount_key#24/1 (work_completion)(&(&sbi->old_work)->work) irq_context: 0 &type->s_umount_key#24/1 &x->wait#23 irq_context: 0 &disk->open_mutex bdev_lock irq_context: 0 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &xa->xa_lock#5 irq_context: 0 sb_lock &obj_hash[i].lock irq_context: 0 sb_lock pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 irq_context: 0 &type->s_umount_key#25/1 fs_reclaim irq_context: 0 &type->s_umount_key#25/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#25/1 shrinker_mutex irq_context: 0 &type->s_umount_key#25/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#25/1 sb_lock irq_context: 0 &type->s_umount_key#25/1 inode_hash_lock irq_context: 0 &type->s_umount_key#25/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#25/1 bdev_lock irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#25/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 &____s->seqcount irq_context: 0 &type->s_umount_key#25/1 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#25/1 lock#4 irq_context: 0 &type->s_umount_key#25/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#25/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#25/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#25/1 &rq->__lock irq_context: 0 &type->s_umount_key#25/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#25/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#25/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#25/1 &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#25/1 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#25/1 lock#5 irq_context: 0 &type->s_umount_key#25/1 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#25/1 &c->lock irq_context: 0 &type->s_umount_key#25/1 crypto_alg_sem irq_context: 0 &type->s_umount_key#25/1 lock#2 irq_context: 0 &type->s_umount_key#25/1 lock#2 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#25/1 lock#2 (work_completion)(work) irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock (wq_completion)mm_percpu_wq irq_context: 0 &type->s_umount_key#25/1 lock#2 &x->wait#10 irq_context: 0 &type->s_umount_key#25/1 lock#2 &rq->__lock irq_context: 0 &type->s_umount_key#25/1 lock#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#5 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#25/1 &x->wait#23 irq_context: 0 tomoyo_ss quarantine_lock irq_context: 0 &type->s_umount_key#26/1 irq_context: 0 &type->s_umount_key#26/1 fs_reclaim irq_context: 0 &type->s_umount_key#26/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26/1 shrinker_mutex irq_context: 0 &type->s_umount_key#26/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#26/1 sb_lock irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#26/1 bdev_lock irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#26/1 &c->lock irq_context: 0 &type->s_umount_key#26/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#26/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#26/1 lock#4 irq_context: 0 &type->s_umount_key#26/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#26/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#26/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#26/1 &rq->__lock irq_context: 0 &type->s_umount_key#26/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#26/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#26/1 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#26/1 lock#5 irq_context: 0 &type->s_umount_key#26/1 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#26/1 crypto_alg_sem irq_context: 0 &type->s_umount_key#26/1 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#9 &c->lock irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#9 &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 percpu_counters_lock irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &____s->seqcount irq_context: hardirq &retval->lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 proc_subdir_lock irq_context: 0 &type->s_umount_key#26/1 proc_inum_ida.xa_lock irq_context: 0 &type->s_umount_key#26/1 proc_subdir_lock irq_context: 0 &type->s_umount_key#26/1 &journal->j_state_lock irq_context: 0 &type->s_umount_key#26/1 kthread_create_lock irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock irq_context: 0 &type->s_umount_key#26/1 &x->wait irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#26/1 &journal->j_wait_done_commit irq_context: 0 &journal->j_wait_done_commit irq_context: 0 &journal->j_wait_done_commit &p->pi_lock irq_context: 0 &type->s_umount_key#26/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit irq_context: 0 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &type->s_umount_key#26/1 &journal->j_state_lock irq_context: 0 &type->s_umount_key#26/1 &p->alloc_lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#26/1 wq_pool_mutex irq_context: 0 &type->s_umount_key#26/1 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->s_umount_key#26/1 &ei->i_es_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &k->list_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &c->lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#26/1 ext4_li_mtx irq_context: 0 &type->s_umount_key#26/1 lock irq_context: 0 &type->s_umount_key#26/1 lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#26/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#26/1 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#26/1 lock kernfs_idr_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 (console_sem).lock irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#26/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: hardirq rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex &x->wait#24 irq_context: 0 rcu_state.barrier_mutex &rq->__lock irq_context: 0 rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &retval->lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &sighand->siglock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &newf->file_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock irq_context: 0 batched_entropy_u16.lock irq_context: 0 batched_entropy_u16.lock crngs.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock ptlock_ptr(ptdesc)#2/1 irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 mapping.invalidate_lock irq_context: 0 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock &____s->seqcount irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 mapping.invalidate_lock lock#4 irq_context: 0 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock &ei->i_es_lock irq_context: 0 mapping.invalidate_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock &c->lock irq_context: 0 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock rcu_read_lock &retval->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 mapping.invalidate_lock &rq->__lock irq_context: 0 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 binfmt_lock irq_context: 0 &fsnotify_mark_srcu irq_context: 0 &xa->xa_lock#9 irq_context: 0 &vma->vm_lock->lock fs_reclaim irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq irq_context: 0 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem irq_context: 0 mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: hardirq &rq->__lock &____s->seqcount irq_context: hardirq &rq->__lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock mapping.invalidate_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &____s->seqcount irq_context: 0 mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &port->mutex irq_context: 0 &tty->ldisc_sem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &port->mutex irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->read_wait irq_context: 0 task_group_lock irq_context: 0 &sighand->siglock &p->pi_lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 lock#4 irq_context: 0 &type->s_umount_key#27/1 irq_context: 0 &type->s_umount_key#27/1 fs_reclaim irq_context: 0 &type->s_umount_key#27/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#27/1 shrinker_mutex irq_context: 0 &type->s_umount_key#27/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#27/1 sb_lock irq_context: 0 &type->s_umount_key#27/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#27/1 &____s->seqcount irq_context: 0 &type->s_umount_key#27/1 &c->lock irq_context: 0 &type->s_umount_key#27/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#27/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#27/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 &type->s_umount_key#27/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 proc_subdir_lock irq_context: 0 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 &p->lock irq_context: 0 &p->lock fs_reclaim irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock pool_lock#2 irq_context: 0 &p->lock &mm->mmap_lock irq_context: 0 &type->s_umount_key#28/1 irq_context: 0 &type->s_umount_key#28/1 fs_reclaim irq_context: 0 &type->s_umount_key#28/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 pool_lock#2 irq_context: 0 &type->s_umount_key#28/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#28/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#28/1 shrinker_mutex irq_context: 0 &type->s_umount_key#28/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#28/1 sb_lock irq_context: 0 &type->s_umount_key#28/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->s_umount_key#28/1 crngs.lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#28/1 &dentry->d_lock irq_context: 0 &root->kernfs_iattr_rwsem irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex irq_context: 0 &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq irq_context: 0 &ent->pde_unload_lock irq_context: 0 &p->lock file_systems_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 &type->s_umount_key#29 irq_context: 0 &type->s_umount_key#29 &x->wait#23 irq_context: 0 &type->s_umount_key#29 shrinker_mutex irq_context: 0 &type->s_umount_key#29 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#29 pool_lock#2 irq_context: 0 &type->s_umount_key#29 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#29 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->s_umount_key#29 rename_lock.seqcount irq_context: 0 &type->s_umount_key#29 &dentry->d_lock irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &type->s_umount_key#29 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#29 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#29 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#29 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#29 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#29 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#29 &obj_hash[i].lock pool_lock irq_context: 0 unnamed_dev_ida.xa_lock irq_context: 0 krc.lock irq_context: 0 &x->wait#25 irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock clock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock irq_context: 0 &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &fs->lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex lock#4 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock cpu_asid_lock irq_context: 0 &sig->wait_chldexit irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock irq_context: 0 tasklist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&ipvs->defense_work)->timer irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&gc_work->dwork)->timer irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &s->s_inode_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->dropentry_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->droppacket_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->securetcp_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &mm->mmap_lock &p->alloc_lock irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 tasklist_lock &sighand->siglock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &c->lock irq_context: 0 tasklist_lock &sighand->siglock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &____s->seqcount#4 irq_context: 0 &prev->lock irq_context: 0 &(&sig->stats_lock)->lock irq_context: 0 &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 &type->i_mutex_dir_key#3 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: hardirq bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 key irq_context: 0 &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 sb_writers#3 irq_context: 0 sb_writers#3 mount_lock irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock irq_context: 0 sb_writers#4 mount_lock irq_context: 0 sb_writers#4 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &wb->list_lock irq_context: 0 sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pid->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq irq_context: 0 &p->alloc_lock &fs->lock &dentry->d_lock irq_context: 0 &p->lock namespace_sem irq_context: 0 &p->lock namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &p->lock namespace_sem rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#30 irq_context: 0 &type->s_umount_key#30 &lru->node[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rsp->gp_wait irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &x->wait#2 irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override task_group_lock irq_context: softirq rcu_callback &x->wait#2 irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sem->waiters irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait pool_lock#2 irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 &p->alloc_lock irq_context: 0 &type->s_umount_key#30 (work_completion)(&sbi->s_sb_upd_work) irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 key#3 irq_context: 0 &type->s_umount_key#30 key#4 irq_context: 0 &type->s_umount_key#30 &sbi->s_error_lock irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 batched_entropy_u8.lock irq_context: 0 &type->s_umount_key#30 kfence_freelist_lock irq_context: 0 &type->s_umount_key#30 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &base->lock irq_context: 0 &type->s_umount_key#30 &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 bit_wait_table + i irq_context: 0 &type->s_umount_key#30 &rq->__lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx irq_context: 0 &type->s_umount_key#30 ext4_li_mtx fs_reclaim irq_context: 0 &type->s_umount_key#30 ext4_li_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &c->lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &____s->seqcount irq_context: 0 &type->s_umount_key#30 ext4_li_mtx pool_lock#2 irq_context: 0 &type->s_umount_key#30 ext4_li_mtx batched_entropy_u16.lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 ext4_li_mtx kthread_create_lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &x->wait irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &rq->__lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 (console_sem).lock irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#30 mount_lock irq_context: 0 &type->s_umount_key#30 mount_lock mount_lock.seqcount irq_context: 0 &type->s_umount_key#30 mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 rcu_read_lock &pid->lock irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rename_lock.seqcount irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &pid->lock irq_context: 0 sb_writers#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &____s->seqcount irq_context: 0 sb_writers#3 &c->lock irq_context: 0 sb_writers#3 pool_lock#2 irq_context: 0 sb_writers#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 jbd2_handle irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle &c->lock irq_context: 0 sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &wb->list_lock irq_context: 0 sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &wb->work_lock irq_context: 0 sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &wb->work_lock &base->lock irq_context: 0 sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal rcu_read_lock init_fs.seq.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_internal pool_lock#2 irq_context: 0 sb_internal &journal->j_state_lock irq_context: 0 sb_internal jbd2_handle irq_context: 0 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_internal jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal &obj_hash[i].lock irq_context: 0 &ei->i_data_sem irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sighand->siglock hrtimer_bases.lock irq_context: 0 &sighand->siglock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sighand->siglock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 file_rwsem irq_context: 0 file_rwsem &ctx->flc_lock irq_context: 0 file_rwsem &ctx->flc_lock &fll->lock irq_context: 0 &ctx->flc_lock irq_context: 0 &sig->cred_guard_mutex tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mount_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#3 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: softirq rcu_callback &cfs_rq->removed.lock irq_context: softirq rcu_callback &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss quarantine_lock irq_context: 0 &sig->cred_guard_mutex &c->lock irq_context: 0 &sig->cred_guard_mutex &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &p->lock &c->lock irq_context: 0 &p->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 rcu_read_lock &p->alloc_lock irq_context: 0 &type->s_umount_key#31/1 irq_context: 0 &type->s_umount_key#31/1 fs_reclaim irq_context: 0 &type->s_umount_key#31/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 pool_lock#2 irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#31/1 shrinker_mutex irq_context: 0 &type->s_umount_key#31/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#31/1 sb_lock irq_context: 0 &type->s_umount_key#31/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#31/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#31/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#31/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 fs_reclaim irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 pool_lock#2 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &p->lock &pcp->lock &zone->lock irq_context: 0 &p->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key/1 fs_reclaim irq_context: 0 &type->s_umount_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key/1 mmu_notifier_invalidate_range_start irq_context: softirq rcu_callback &rsp->gp_wait irq_context: 0 &type->s_umount_key#32 irq_context: 0 &type->s_umount_key#32 sb_lock irq_context: 0 &type->s_umount_key#32 fs_reclaim irq_context: 0 &type->s_umount_key#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 pool_lock#2 irq_context: 0 &type->s_umount_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#32 &lru->node[i].lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#32 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key#33 irq_context: 0 &type->s_umount_key#33 sb_lock irq_context: 0 &type->s_umount_key#33 &dentry->d_lock irq_context: 0 &type->s_umount_key#34 irq_context: 0 &type->s_umount_key#34 sb_lock irq_context: 0 &type->s_umount_key#34 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 &type->s_umount_key#35/1 irq_context: 0 &type->s_umount_key#35/1 fs_reclaim irq_context: 0 &type->s_umount_key#35/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 pool_lock#2 irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#35/1 shrinker_mutex irq_context: 0 &type->s_umount_key#35/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#35/1 sb_lock irq_context: 0 &type->s_umount_key#35/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#35/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#26 irq_context: 0 &type->s_umount_key#35/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#35/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 &type->s_umount_key#35/1 &c->lock irq_context: 0 &type->s_umount_key#35/1 &____s->seqcount irq_context: 0 &type->s_umount_key#35/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key#36 irq_context: 0 &type->s_umount_key#36 sb_lock irq_context: 0 &type->s_umount_key#36 &dentry->d_lock irq_context: 0 redirect_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &port->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &tty->write_wait irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &tty->write_wait &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &port->lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->files_lock irq_context: 0 &tty->ldisc_sem &tty->write_wait irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 &type->s_umount_key#37/1 irq_context: 0 &type->s_umount_key#37/1 fs_reclaim irq_context: 0 &type->s_umount_key#37/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37/1 pool_lock#2 irq_context: 0 &type->s_umount_key#37/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#37/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#37/1 shrinker_mutex irq_context: 0 &type->s_umount_key#37/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#37/1 &c->lock irq_context: 0 &type->s_umount_key#37/1 &____s->seqcount irq_context: 0 &type->s_umount_key#37/1 sb_lock irq_context: 0 &type->s_umount_key#37/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#37/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37/1 &sb->s_type->i_lock_key#27 irq_context: 0 &type->s_umount_key#37/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#37/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#37/1 &sb->s_type->i_lock_key#27 &dentry->d_lock irq_context: 0 &type->s_umount_key#37/1 fuse_mutex irq_context: 0 &type->s_umount_key#37/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#38/1 irq_context: 0 &type->s_umount_key#38/1 fs_reclaim irq_context: 0 &type->s_umount_key#38/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 pool_lock#2 irq_context: 0 &type->s_umount_key#38/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#38/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#38/1 shrinker_mutex irq_context: 0 &type->s_umount_key#38/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#38/1 sb_lock irq_context: 0 &type->s_umount_key#38/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#38/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_lock_key#28 irq_context: 0 &type->s_umount_key#38/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#38/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_lock_key#28 &dentry->d_lock irq_context: 0 &type->s_umount_key#38/1 pstore_sb_lock irq_context: 0 &type->s_umount_key#38/1 pstore_sb_lock &sb->s_type->i_mutex_key#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 fs_reclaim irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &zone->lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex (efivars_lock).lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex fs_reclaim irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &c->lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex (efi_runtime_lock).lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &x->wait#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_node_0 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#39/1 irq_context: 0 &type->s_umount_key#39/1 fs_reclaim irq_context: 0 &type->s_umount_key#39/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#39/1 shrinker_mutex irq_context: 0 &type->s_umount_key#39/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#39/1 sb_lock irq_context: 0 &type->s_umount_key#39/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#39/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 irq_context: 0 &type->s_umount_key#39/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#39/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 &dentry->d_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock (kmod_concurrent_max).lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &x->wait#17 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uts_sem irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock key irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock pcpu_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock percpu_counters_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock running_helpers_waitq.lock irq_context: 0 &type->s_umount_key#39/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 irq_context: 0 &type->s_umount_key#14 sb_lock irq_context: 0 &type->s_umount_key#14 fs_reclaim irq_context: 0 &type->s_umount_key#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#14 pool_lock#2 irq_context: 0 &type->s_umount_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &lru->node[i].lock irq_context: 0 &type->s_umount_key#14 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key irq_context: 0 &type->i_mutex_dir_key#5 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq irq_context: 0 sb_writers#5 irq_context: 0 sb_writers#5 mount_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock pool_lock#2 irq_context: softirq (&cb->timer) irq_context: softirq (&cb->timer) &obj_hash[i].lock irq_context: softirq (&cb->timer) &base->lock irq_context: softirq (&cb->timer) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &wb->list_lock irq_context: 0 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#3 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#3 pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &mm->mmap_lock sb_writers#3 &obj_hash[i].lock irq_context: 0 &f->f_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 uts_sem irq_context: 0 uts_sem hostname_poll.wait.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &f->f_pos_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &mm->mmap_lock irq_context: 0 &fs->lock &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mm->page_table_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock ptlock_ptr(ptdesc)#2/1 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock#2 irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xattrs->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#9 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &info->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 irq_context: 0 key#5 irq_context: 0 &p->alloc_lock &x->wait#25 irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss key irq_context: 0 tomoyo_ss pcpu_lock irq_context: 0 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &u->iolock irq_context: 0 &u->iolock rlock-AF_UNIX irq_context: 0 &ei->socket.wq.wait irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &obj_hash[i].lock irq_context: 0 &sighand->siglock pool_lock#2 irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &stopper->lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &x->wait#8 irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 &c->lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 &bsd_socket_locks[i] irq_context: 0 sb_writers tk_core.seq.seqcount irq_context: 0 sb_writers &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &wb->list_lock irq_context: 0 sb_writers &wb->list_lock &sb->s_type->i_lock_key#5 irq_context: 0 &u->lock irq_context: 0 &u->lock &u->lock/1 irq_context: 0 &u->lock rlock-AF_UNIX irq_context: 0 rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock irq_context: 0 &u->iolock &obj_hash[i].lock irq_context: 0 &u->iolock pool_lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &wb->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 syslog_lock irq_context: 0 &u->lock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#14 irq_context: 0 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &pipe->mutex/1 irq_context: 0 &pipe->rd_wait irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&cb->timer) tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 key#6 irq_context: 0 &u->iolock quarantine_lock irq_context: 0 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 &u->lock clock-AF_UNIX irq_context: 0 &u->peer_wait irq_context: 0 rlock-AF_UNIX irq_context: 0 &pipe->mutex/1 &pipe->rd_wait irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &pipe->wr_wait irq_context: 0 &pipe->mutex/1 fs_reclaim irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock irq_context: 0 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#6 tk_core.seq.seqcount irq_context: 0 sb_writers#6 mount_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &rq->__lock irq_context: 0 &pipe->mutex/1 &lock->wait_lock irq_context: 0 &pipe->mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &lock->wait_lock irq_context: 0 sb_writers#3 batched_entropy_u8.lock irq_context: 0 sb_writers#3 batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#3 kfence_freelist_lock irq_context: 0 sb_writers#3 &meta->lock irq_context: 0 sb_writers#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#4 sysctl_lock irq_context: 0 sb_writers#4 &dentry->d_lock irq_context: 0 sb_writers#4 tomoyo_ss irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#4 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#4 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#4 fs_reclaim irq_context: 0 sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 pool_lock#2 irq_context: 0 sb_writers#4 &mm->mmap_lock irq_context: 0 sb_writers#4 &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock pool_lock#2 irq_context: 0 nl_table_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK slock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 slock-AF_NETLINK irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 &u->lock &sk->sk_peer_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 mapping.invalidate_lock &ei->i_es_lock key#7 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: hardirq log_wait.lock &p->pi_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xattrs->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 &vma->vm_lock->lock &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 oom_adj_mutex irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &p->alloc_lock irq_context: 0 sb_writers#4 oom_adj_mutex &p->alloc_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 &group->mark_mutex irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex fs_reclaim irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &group->mark_mutex &____s->seqcount irq_context: 0 &group->mark_mutex &c->lock irq_context: 0 &group->mark_mutex pool_lock#2 irq_context: 0 &group->mark_mutex lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &group->mark_mutex ucounts_lock irq_context: 0 &group->mark_mutex &mark->lock irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu &conn->lock irq_context: 0 &group->mark_mutex &mark->lock &conn->lock irq_context: 0 &group->mark_mutex &conn->lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock pool_lock#2 irq_context: 0 &sk->sk_peer_lock irq_context: 0 &ep->mtx irq_context: 0 epnested_mutex irq_context: 0 epnested_mutex &ep->mtx irq_context: 0 epnested_mutex &ep->mtx fs_reclaim irq_context: 0 epnested_mutex &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 epnested_mutex &ep->mtx &____s->seqcount irq_context: 0 epnested_mutex &ep->mtx &c->lock irq_context: 0 epnested_mutex &ep->mtx pool_lock#2 irq_context: 0 epnested_mutex &ep->mtx &f->f_lock irq_context: 0 epnested_mutex &ep->mtx &ei->socket.wq.wait irq_context: 0 epnested_mutex &ep->mtx &ep->lock irq_context: 0 epnested_mutex rcu_read_lock &f->f_lock irq_context: 0 &ep->mtx fs_reclaim irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx &f->f_lock irq_context: 0 &ep->mtx pool_lock#2 irq_context: 0 &ep->mtx &group->notification_waitq irq_context: 0 &ep->mtx &group->notification_lock irq_context: 0 &ep->mtx &ep->lock irq_context: 0 &ep->mtx &sighand->signalfd_wqh irq_context: 0 &ep->mtx &sighand->siglock irq_context: 0 &ep->mtx &ei->socket.wq.wait irq_context: 0 &ep->lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &c->lock irq_context: 0 remove_cache_srcu &n->list_lock irq_context: 0 remove_cache_srcu &obj_hash[i].lock irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tomoyo_ss &n->list_lock irq_context: 0 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 vmap_area_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tomoyo_ss remove_cache_srcu irq_context: 0 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&net->can.stattimer) irq_context: softirq (&net->can.stattimer) &obj_hash[i].lock irq_context: softirq (&net->can.stattimer) &base->lock irq_context: softirq (&net->can.stattimer) &base->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &sb->s_type->i_lock_key#24 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: hardirq &dev->power.lock hrtimer_bases.lock irq_context: hardirq &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_callback &pcp->lock &zone->lock irq_context: softirq rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: 0 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#4 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#4 kfence_freelist_lock irq_context: 0 &type->i_mutex_dir_key#4 &meta->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 &type->i_mutex_dir_key#4 quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&krcp->monitor_work)->timer irq_context: softirq &(&tbl->managed_work)->timer irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 swap_lock irq_context: 0 sb_writers#7 irq_context: 0 sb_writers#7 mount_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 iattr_mutex irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 kn->active fs_reclaim irq_context: 0 kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active pool_lock#2 irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 sb_writers#7 fs_reclaim irq_context: 0 sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 pool_lock#2 irq_context: 0 sb_writers#7 &mm->mmap_lock irq_context: 0 sb_writers#7 &of->mutex irq_context: 0 sb_writers#7 &of->mutex kn->active &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active &obj_hash[i].lock irq_context: 0 sb_writers#7 &obj_hash[i].lock irq_context: 0 &ep->mtx &mm->mmap_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rlock-AF_NETLINK irq_context: 0 &nlk->wait irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 fs_reclaim irq_context: 0 kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#2 &c->lock irq_context: 0 kn->active#2 &____s->seqcount irq_context: 0 kn->active#2 &n->list_lock irq_context: 0 kn->active#2 &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 kn->active#2 &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active &____s->seqcount irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active &c->lock irq_context: 0 kn->active &n->list_lock irq_context: 0 kn->active &n->list_lock &c->lock irq_context: 0 kn->active &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu irq_context: 0 kn->active#2 remove_cache_srcu irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#2 remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &c->lock irq_context: 0 kn->active &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 quarantine_lock irq_context: 0 sb_writers#7 &n->list_lock irq_context: 0 sb_writers#7 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#2 remove_cache_srcu &c->lock irq_context: 0 kn->active#2 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 remove_cache_srcu irq_context: 0 sb_writers#7 remove_cache_srcu quarantine_lock irq_context: 0 &ep->mtx &c->lock irq_context: 0 &ep->mtx &____s->seqcount irq_context: 0 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 &rq->__lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active quarantine_lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex quarantine_lock irq_context: 0 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#2 remove_cache_srcu &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock irq_context: softirq (&vblank->disable_timer) irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#2 batched_entropy_u8.lock irq_context: 0 kn->active#2 kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 kn->active#3 fs_reclaim irq_context: 0 kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#3 &c->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &obj_hash[i].lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#3 &pcp->lock &zone->lock irq_context: 0 kn->active#3 &____s->seqcount irq_context: 0 kn->active#3 &n->list_lock irq_context: 0 kn->active#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &pcp->lock &zone->lock irq_context: 0 kn->active#3 &rq->__lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#3 remove_cache_srcu irq_context: 0 kn->active#3 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#3 remove_cache_srcu &c->lock irq_context: 0 kn->active#3 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&q->timeout) irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 (wq_completion)kblockd irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#3 batched_entropy_u8.lock irq_context: 0 kn->active#3 kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &rq->__lock irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &wb->list_lock irq_context: 0 sb_writers#7 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#4 fs_reclaim irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &c->lock irq_context: 0 kn->active#4 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &pcp->lock &zone->lock irq_context: 0 kn->active#4 remove_cache_srcu irq_context: 0 kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#4 &n->list_lock irq_context: 0 kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &device->physical_node_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active#4 remove_cache_srcu &c->lock irq_context: 0 kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 udc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fw_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &rq->__lock irq_context: 0 sb_writers#7 batched_entropy_u8.lock irq_context: 0 sb_writers#7 kfence_freelist_lock irq_context: 0 sb_writers#7 &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 quarantine_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#4 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rfkill->lock irq_context: 0 kn->active#4 &pcp->lock &zone->lock irq_context: 0 kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#4 batched_entropy_u8.lock irq_context: 0 kn->active#4 kfence_freelist_lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 tomoyo_ss batched_entropy_u8.lock irq_context: 0 tomoyo_ss kfence_freelist_lock irq_context: 0 tomoyo_ss &meta->lock irq_context: 0 &u->bindlock irq_context: 0 &u->bindlock fs_reclaim irq_context: 0 &u->bindlock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->bindlock pool_lock#2 irq_context: 0 &u->bindlock batched_entropy_u32.lock irq_context: 0 &u->bindlock &net->unx.table.locks[i] irq_context: 0 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock irq_context: 0 &u->lock &u->lock/1 &dentry->d_lock irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock &sk->sk_peer_lock/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock/1 irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &dentry->d_lock irq_context: 0 &ep->mtx &rq->__lock irq_context: 0 &ep->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex &pcp->lock &zone->lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &group->notification_waitq irq_context: 0 &group->notification_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock pool_lock#2 irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &p->pi_lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 &u->iolock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#30 sb_writers#3 lock#4 irq_context: 0 &type->s_umount_key#30 sb_writers#3 pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &base->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 bit_wait_table + i irq_context: 0 &type->s_umount_key#30 sb_writers#3 &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 sb_writers#3 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &x->wait#26 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &__ctx->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &__ctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &retval->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &x->wait#26 irq_context: hardirq &x->wait#26 &p->pi_lock irq_context: hardirq &x->wait#26 &p->pi_lock &rq->__lock irq_context: hardirq &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem (&timer.timer) irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &q->requeue_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &nvmeq->sq_lock irq_context: hardirq &fq->mq_flush_lock irq_context: hardirq &fq->mq_flush_lock tk_core.seq.seqcount irq_context: hardirq &fq->mq_flush_lock &x->wait#26 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &type->s_umount_key#30 sb_writers#3 &obj_hash[i].lock irq_context: 0 tomoyo_ss batched_entropy_u8.lock crngs.lock irq_context: 0 tomoyo_ss batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &n->list_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &____s->seqcount irq_context: softirq &(&ovs_net->masks_rebalance)->timer irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &p->pi_lock irq_context: 0 kn->active#5 fs_reclaim irq_context: 0 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#5 &c->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex irq_context: 0 &p->lock &of->mutex kn->active#5 param_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &dentry->d_lock irq_context: 0 sb_writers#7 tomoyo_ss irq_context: 0 sb_writers#7 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#5 fs_reclaim irq_context: 0 sb_writers#7 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 tomoyo_ss &c->lock irq_context: 0 sb_writers#7 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#7 iattr_mutex irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &pcp->lock &zone->lock irq_context: 0 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &____s->seqcount irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &c->lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &sem->wait_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#5 param_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#5 param_lock disk_events_mutex irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#4 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &n->list_lock irq_context: 0 &p->lock &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#4 &device->physical_node_lock irq_context: 0 &p->lock remove_cache_srcu irq_context: 0 &p->lock remove_cache_srcu quarantine_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu irq_context: 0 &vma->vm_lock->lock remove_cache_srcu quarantine_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &n->list_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 kn->active#6 fs_reclaim irq_context: 0 kn->active#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#7 fs_reclaim irq_context: 0 kn->active#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock &c->lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 fs_reclaim irq_context: 0 kn->active#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#9 fs_reclaim irq_context: 0 kn->active#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 kn->active#10 fs_reclaim irq_context: 0 kn->active#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 kn->active#11 fs_reclaim irq_context: 0 kn->active#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#11 dev_base_lock irq_context: 0 kn->active#12 fs_reclaim irq_context: 0 kn->active#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#12 dev_base_lock irq_context: 0 kn->active#13 fs_reclaim irq_context: 0 kn->active#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#14 fs_reclaim irq_context: 0 kn->active#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#14 dev_base_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#15 fs_reclaim irq_context: 0 kn->active#15 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#15 remove_cache_srcu irq_context: 0 kn->active#15 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#15 remove_cache_srcu &c->lock irq_context: 0 kn->active#15 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#15 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#15 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#15 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#15 dev_base_lock irq_context: 0 kn->active#6 &c->lock irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#10 &c->lock irq_context: 0 kn->active#10 &n->list_lock irq_context: 0 kn->active#10 &n->list_lock &c->lock irq_context: 0 kn->active#16 fs_reclaim irq_context: 0 kn->active#16 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#17 fs_reclaim irq_context: 0 kn->active#17 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &c->lock irq_context: 0 kn->active#18 fs_reclaim irq_context: 0 kn->active#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 fs_reclaim irq_context: 0 kn->active#19 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#20 fs_reclaim irq_context: 0 kn->active#20 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock irq_context: 0 rcu_read_lock &____s->seqcount#6 irq_context: 0 kn->active#21 fs_reclaim irq_context: 0 kn->active#21 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#21 dev_base_lock irq_context: 0 &sb->s_type->i_mutex_key#15 irq_context: 0 mapping.invalidate_lock#2 irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock#2 &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &c->lock irq_context: 0 mapping.invalidate_lock#2 lock#4 irq_context: 0 mapping.invalidate_lock#2 pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &nvmeq->sq_lock irq_context: 0 kn->active#22 fs_reclaim irq_context: 0 kn->active#22 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &of->mutex irq_context: 0 &of->mutex kn->active#22 &dev->power.lock irq_context: 0 &of->mutex kn->active#22 pci_lock irq_context: 0 kn->active#23 fs_reclaim irq_context: 0 kn->active#23 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#24 fs_reclaim irq_context: 0 kn->active#24 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#24 &c->lock irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#17 &c->lock irq_context: 0 kn->active#17 &____s->seqcount irq_context: 0 kn->active#25 fs_reclaim irq_context: 0 kn->active#25 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 slock-AF_INET/1 irq_context: 0 rtnl_mutex devnet_rename_sem irq_context: 0 rtnl_mutex devnet_rename_sem (console_sem).lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock#2 lock#4 &lruvec->lru_lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock irq_context: 0 kn->active#26 &rq->__lock irq_context: 0 kn->active#26 fs_reclaim irq_context: 0 kn->active#26 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#26 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#26 fs_reclaim &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#26 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#26 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#26 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#26 &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &k->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex devnet_rename_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex &x->wait#2 irq_context: 0 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#27 fs_reclaim irq_context: 0 kn->active#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 fs_reclaim irq_context: 0 kn->active#28 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#29 fs_reclaim irq_context: 0 kn->active#29 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#30 fs_reclaim irq_context: 0 kn->active#30 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#31 fs_reclaim irq_context: 0 kn->active#31 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#31 &c->lock irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#32 fs_reclaim irq_context: 0 kn->active#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#33 fs_reclaim irq_context: 0 kn->active#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#33 &c->lock irq_context: 0 kn->active#33 &n->list_lock irq_context: 0 kn->active#33 &n->list_lock &c->lock irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss quarantine_lock irq_context: 0 &type->i_mutex_dir_key#2 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &evdev->client_lock irq_context: 0 &evdev->mutex irq_context: 0 &evdev->mutex &dev->mutex#2 irq_context: 0 mapping.invalidate_lock#2 &c->lock irq_context: 0 sb_writers &sem->wait_lock irq_context: 0 sb_writers &p->pi_lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock &c->lock irq_context: 0 kn->active#4 &rq->__lock irq_context: 0 kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 mapping.invalidate_lock#2 &pcp->lock &zone->lock irq_context: 0 mapping.invalidate_lock#2 &n->list_lock irq_context: 0 mapping.invalidate_lock#2 &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#4 fw_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &retval->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &____s->seqcount irq_context: 0 lock#4 &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 kn->active#34 fs_reclaim irq_context: 0 kn->active#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &c->lock irq_context: 0 kn->active#11 &c->lock irq_context: 0 &p->lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &n->list_lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 rtnl_mutex sysctl_lock krc.lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &nft_net->commit_mutex irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex &ent->pde_unload_lock irq_context: 0 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex target_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET#2 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET irq_context: 0 &evdev->mutex &dev->mutex#2 &obj_hash[i].lock irq_context: 0 &evdev->mutex &dev->mutex#2 &x->wait#2 irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &info->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xa->xa_lock#9 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sem->wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 kn->active#18 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &rq->__lock irq_context: 0 sb_writers &rq->__lock irq_context: 0 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#19 &c->lock irq_context: 0 kn->active#19 &n->list_lock irq_context: 0 kn->active#19 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#35 fs_reclaim irq_context: 0 kn->active#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#35 &c->lock irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 remove_cache_srcu irq_context: 0 kn->active#20 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#20 remove_cache_srcu &c->lock irq_context: 0 kn->active#20 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#20 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#20 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#20 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#16 &c->lock irq_context: 0 kn->active#16 &n->list_lock irq_context: 0 kn->active#16 &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &n->list_lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#27 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#32 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock.wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock irq_context: 0 kn->active#27 &n->list_lock irq_context: 0 kn->active#27 &n->list_lock &c->lock irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#28 &c->lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &eli->li_list_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx pool_lock#2 irq_context: 0 ext4_li_mtx irq_context: 0 ext4_li_mtx &eli->li_list_mtx irq_context: 0 ext4_li_mtx &obj_hash[i].lock irq_context: 0 ext4_li_mtx pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#30 &c->lock irq_context: 0 kn->active#31 &n->list_lock irq_context: 0 kn->active#31 &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 &rq->__lock irq_context: 0 kn->active#27 remove_cache_srcu irq_context: 0 kn->active#27 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#36 fs_reclaim irq_context: 0 kn->active#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 fs_reclaim irq_context: 0 kn->active#37 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 &c->lock irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock remove_cache_srcu &c->lock irq_context: 0 &p->lock remove_cache_srcu &n->list_lock irq_context: 0 &p->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &p->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#27 remove_cache_srcu &c->lock irq_context: 0 kn->active#27 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#27 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#27 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#27 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#29 &c->lock irq_context: 0 kn->active#29 &n->list_lock irq_context: 0 kn->active#29 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#32 &n->list_lock irq_context: 0 kn->active#32 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &c->lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss quarantine_lock irq_context: 0 kn->active#28 &n->list_lock irq_context: 0 kn->active#28 &n->list_lock &c->lock irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#32 remove_cache_srcu irq_context: 0 kn->active#30 remove_cache_srcu irq_context: 0 kn->active#32 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#30 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#32 remove_cache_srcu &c->lock irq_context: 0 kn->active#32 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#32 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#32 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#32 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 kn->active#30 &n->list_lock irq_context: 0 kn->active#30 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#30 &pcp->lock &zone->lock irq_context: 0 kn->active#30 &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &pipe->rd_wait irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx key#10 irq_context: 0 &sig->cred_guard_mutex &n->list_lock irq_context: 0 &sig->cred_guard_mutex &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 videodev_lock irq_context: 0 &dev_instance->mutex irq_context: 0 &dev_instance->mutex fs_reclaim irq_context: 0 &dev_instance->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev_instance->mutex pool_lock#2 irq_context: 0 &dev_instance->mutex vicodec_core:1851:(hdl)->_lock irq_context: 0 &dev_instance->mutex &vdev->fh_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock irq_context: 0 &mdev->req_queue_mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex &mdev->graph_mutex irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex quarantine_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &pipe->rd_wait &ep->lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &folio_wait_table[i] irq_context: 0 &vma->vm_lock->lock &folio_wait_table[i] &p->pi_lock irq_context: 0 &vma->vm_lock->lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex key#5 irq_context: 0 &dev->dev_mutex irq_context: 0 &dev->dev_mutex fs_reclaim irq_context: 0 &dev->dev_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->dev_mutex &c->lock irq_context: 0 &dev->dev_mutex pool_lock#2 irq_context: 0 &dev->dev_mutex &____s->seqcount irq_context: 0 &dev->dev_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &dev->dev_mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex pool_lock#2 irq_context: 0 &ep->mtx rcu_read_lock &pipe->rd_wait irq_context: 0 &ep->mtx &obj_hash[i].lock irq_context: 0 &sighand->signalfd_wqh irq_context: 0 &vdev->fh_lock irq_context: 0 &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &vdev->fh_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &q->mmap_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#38 fs_reclaim irq_context: 0 kn->active#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock kfence_freelist_lock irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 &sighand->siglock &meta->lock irq_context: 0 &sighand->siglock kfence_freelist_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &mm->page_table_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &p->lock &rq->__lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock irq_context: 0 &dev_instance->mutex &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 fh->state->lock irq_context: 0 rcu_read_lock key irq_context: 0 rcu_read_lock pcpu_lock irq_context: 0 rcu_read_lock percpu_counters_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#38 &c->lock irq_context: 0 kn->active#38 &____s->seqcount irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#38 &pcp->lock &zone->lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &rq->__lock cpu_asid_lock irq_context: 0 cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &vma->vm_lock->lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem quarantine_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock quarantine_lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock pool_lock#2 irq_context: 0 &vcapture->lock irq_context: 0 &vcapture->lock &q->done_wq irq_context: 0 &vcapture->lock &q->mmap_lock irq_context: 0 &mdev->graph_mutex irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 remove_cache_srcu irq_context: 0 sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback quarantine_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock fs_reclaim &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &lo->lo_mutex irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &disk->open_mutex &lo->lo_mutex irq_context: 0 &group->mark_mutex &rq->__lock irq_context: 0 &group->mark_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &bdev->bd_size_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &q->queue_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &x->wait#3 irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &group->mark_mutex remove_cache_srcu irq_context: 0 &group->mark_mutex remove_cache_srcu quarantine_lock irq_context: 0 &group->mark_mutex remove_cache_srcu &c->lock irq_context: 0 &group->mark_mutex remove_cache_srcu &n->list_lock irq_context: 0 &group->mark_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &group->mark_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &group->mark_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#4 &rfkill->lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &rq->__lock irq_context: 0 kn->active#13 &c->lock irq_context: 0 kn->active#13 &pcp->lock &zone->lock irq_context: 0 kn->active#13 &____s->seqcount irq_context: 0 kn->active#11 &pcp->lock &zone->lock irq_context: 0 kn->active#11 &____s->seqcount irq_context: 0 kn->active#15 &c->lock irq_context: 0 kn->active#15 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock irq_context: 0 &disk->open_mutex &new->lock irq_context: 0 &disk->open_mutex &new->lock &mtdblk->cache_mutex irq_context: 0 kn->active#39 fs_reclaim irq_context: 0 kn->active#39 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mtd->master.chrdev_lock irq_context: 0 &mtd->master.chrdev_lock &mm->mmap_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#4 fs_reclaim &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &p->lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &p->lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 rename_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &wq irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock pool_lock#2 irq_context: 0 &dentry->d_lock &lru->node[i].lock irq_context: 0 kn->active#40 fs_reclaim irq_context: 0 kn->active#40 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#21 &c->lock irq_context: 0 kn->active#23 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &info->lock irq_context: 0 sb_writers#5 &obj_hash[i].lock irq_context: 0 sb_writers#5 pool_lock#2 irq_context: 0 sb_writers#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &xa->xa_lock#9 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu irq_context: 0 &mark->lock irq_context: 0 &group->inotify_data.idr_lock irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock irq_context: 0 &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &conn->lock irq_context: 0 destroy_lock irq_context: 0 fs/notify/mark.c:89 irq_context: 0 (wq_completion)events_unbound connector_reaper_work irq_context: 0 (reaper_work).work irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work irq_context: 0 (wq_completion)events_unbound (reaper_work).work destroy_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound (reaper_work).work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &x->wait#3 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work destroy_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound connector_reaper_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&wb->dwork)->timer irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &x->wait#3 irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &p->sequence irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (&journal->j_commit_timer) irq_context: 0 &journal->j_checkpoint_mutex irq_context: 0 &journal->j_checkpoint_mutex mmu_notifier_invalidate_range_start irq_context: 0 &journal->j_checkpoint_mutex pool_lock#2 irq_context: 0 &journal->j_checkpoint_mutex tk_core.seq.seqcount irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 &journal->j_checkpoint_mutex bit_wait_table + i irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_updates irq_context: 0 &journal->j_list_lock irq_context: 0 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &ei->i_es_lock irq_context: 0 &mapping->i_private_lock irq_context: 0 &ret->b_state_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock irq_context: 0 &ei->i_es_lock key#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 rcu_read_lock &retval->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &__ctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &__ctx->lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &memcg->move_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_md_lock irq_context: 0 &journal->j_fc_wait irq_context: 0 &journal->j_history_lock irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#3 batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock sb_writers#3 kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &meta->lock irq_context: 0 rtnl_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &dir->lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket rhashtable_bucket/1 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETLINK irq_context: 0 &sb->s_type->i_mutex_key#10 &nlk->wait irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem hwsim_radio_lock irq_context: 0 rtnl_mutex _xmit_LOOPBACK irq_context: 0 rtnl_mutex netpoll_srcu irq_context: 0 rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex &im->lock irq_context: 0 rtnl_mutex fib_info_lock irq_context: 0 rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock#2 irq_context: 0 rtnl_mutex cbs_list_lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &idev->mc_lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex &ifa->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex &n->list_lock irq_context: 0 rtnl_mutex &n->list_lock &c->lock irq_context: softirq rcu_callback &dir->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: softirq rcu_callback &dir->lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET6 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#5 tomoyo_ss irq_context: 0 sb_writers#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &xattrs->lock irq_context: 0 &u->lock/1 irq_context: 0 userns_state_mutex irq_context: 0 vmap_purge_lock free_vmap_area_lock quarantine_lock irq_context: 0 mapping.invalidate_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mapping->i_mmap_rwsem irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#5 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#5 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &info->lock irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 irq_context: 0 &sb->s_type->i_lock_key#4 irq_context: 0 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 sk_lock-AF_UNIX irq_context: 0 sk_lock-AF_UNIX slock-AF_UNIX irq_context: 0 slock-AF_UNIX irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem key irq_context: 0 cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sk_lock-AF_INET irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 irq_context: 0 slock-AF_INET#2 irq_context: 0 sk_lock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 slock-AF_INET6 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock key#7 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &n->list_lock &c->lock irq_context: 0 free_vmap_area_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &table->hash[i].lock irq_context: 0 sk_lock-AF_INET &table->hash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_NETLINK &mm->mmap_lock irq_context: 0 sk_lock-AF_NETLINK fs_reclaim irq_context: 0 sk_lock-AF_NETLINK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_NETLINK pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK &____s->seqcount irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_NETLINK &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK &c->lock irq_context: 0 sk_lock-AF_NETLINK batched_entropy_u32.lock irq_context: 0 sk_lock-AF_NETLINK vmap_purge_lock irq_context: 0 sk_lock-AF_NETLINK vmap_purge_lock purge_vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK &fp->aux->used_maps_mutex irq_context: 0 kn->active#41 fs_reclaim irq_context: 0 kn->active#41 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock &ifa->lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq#2 irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock irq_context: 0 cb_lock fs_reclaim irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &c->lock irq_context: 0 cb_lock pool_lock#2 irq_context: 0 cb_lock rlock-AF_NETLINK irq_context: 0 cb_lock rtnl_mutex irq_context: 0 cb_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &c->lock irq_context: 0 cb_lock genl_mutex &____s->seqcount irq_context: 0 dev_addr_sem irq_context: 0 cb_lock &pcp->lock &zone->lock irq_context: 0 cb_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock &c->lock irq_context: 0 cb_lock &n->list_lock irq_context: 0 cb_lock &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->alloc_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 sb_writers#5 fs_reclaim irq_context: 0 sb_writers#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 lock#4 irq_context: 0 sb_writers#5 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 lock#5 irq_context: 0 sb_writers#5 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 vmap_area_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: softirq (&net->sctp.addr_wq_timer) irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &anon_vma->rwsem ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &meta->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#5 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &dev_addr_list_lock_key pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &pnettable->lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex irq_context: 0 rtnl_mutex napi_hash_lock irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex remove_cache_srcu irq_context: 0 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex x25_neigh_list_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner irq_context: 0 rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &u->lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex _xmit_ETHER irq_context: 0 rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex _xmit_SLIP irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: softirq (&eql->timer) irq_context: softirq (&eql->timer) &eql->queue.lock irq_context: softirq (&eql->timer) &obj_hash[i].lock irq_context: softirq (&eql->timer) &base->lock irq_context: softirq (&eql->timer) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock irq_context: 0 rtnl_mutex free_vmap_area_lock irq_context: 0 rtnl_mutex vmap_area_lock irq_context: 0 rtnl_mutex init_mm.page_table_lock irq_context: 0 rtnl_mutex &cma->lock irq_context: 0 rtnl_mutex cma_mutex irq_context: 0 rtnl_mutex cma_mutex &zone->lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex cma_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 irq_context: 0 rtnl_mutex cma_mutex lock#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex lock#2 &rq->__lock irq_context: 0 rtnl_mutex cma_mutex lock#2 (work_completion)(work) irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock (wq_completion)mm_percpu_wq irq_context: 0 rtnl_mutex cma_mutex lock#2 &x->wait#10 irq_context: 0 rtnl_mutex cma_mutex lock#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex fs_reclaim irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex patch_lock irq_context: 0 rtnl_mutex &zone->lock irq_context: 0 rtnl_mutex rcu_node_0 irq_context: hardirq &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) (console_sem).lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &base->lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &dir->lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex cbs_list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex &rfkill->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &data->mutex irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->filter_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_VOID irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nl_table_wait.lock &p->pi_lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &u->lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock &____s->seqcount irq_context: 0 &u->iolock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mapping.invalidate_lock lock#4 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 batched_entropy_u16.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex _xmit_X25 irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex lapb_list_lock pool_lock#2 irq_context: 0 rtnl_mutex lapb_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapbeth->up_lock irq_context: 0 rtnl_mutex &lapb->lock irq_context: 0 rtnl_mutex &lapb->lock pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &h->lhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &mm->mmap_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &mapping->i_private_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: softirq rcu_callback &ul->lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &tty->legacy_mutex &f->f_lock irq_context: 0 &tty->legacy_mutex &f->f_lock fasync_lock irq_context: 0 &tty->legacy_mutex &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex tasklist_lock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock &tty->ctrl.lock irq_context: 0 rcu_read_lock &tty->ctrl.lock irq_context: 0 &tty->ctrl.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 &port_lock_key irq_context: 0 &buf->lock irq_context: 0 &tty->ldisc_sem rcu_read_lock &tty->ctrl.lock irq_context: 0 &tty->ldisc_sem &port_lock_key irq_context: 0 &tty->ldisc_sem &port->lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->flow.lock irq_context: softirq rcu_callback rcu_read_lock rt6_exception_lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex lapb_list_lock &c->lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem &tty->read_wait irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock (work_completion)(&buf->work) irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &lapb->lock &c->lock irq_context: softirq &tx->clean_lock irq_context: softirq &tx->clean_lock &obj_hash[i].lock irq_context: softirq &tx->clean_lock pool_lock#2 irq_context: 0 sb_writers#3 &rq->__lock irq_context: softirq &(&idev->mc_dad_work)->timer irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock irq_context: 0 tasklist_lock &sighand->siglock batched_entropy_u8.lock crngs.lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &pcp->lock &zone->lock irq_context: 0 rtnl_mutex class irq_context: 0 rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&idev->mc_ifc_work)->timer irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: softirq rcu_callback &ul->lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &____s->seqcount irq_context: 0 lock pidmap_lock &n->list_lock irq_context: 0 lock pidmap_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 fs_reclaim irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock irq_context: 0 sk_lock-AF_INET6 once_lock irq_context: 0 sk_lock-AF_INET6 once_lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq net/core/link_watch.c:31 irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock noop_qdisc.q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex class irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sb_writers#5 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock krc.lock irq_context: 0 rtnl_mutex &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rt6_exception_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &dir->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &obj_hash[i].lock irq_context: 0 &net->packet.sklist_lock irq_context: 0 sk_lock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET &po->bind_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock ptype_lock irq_context: 0 sk_lock-AF_PACKET &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &x->wait#2 irq_context: 0 sk_lock-AF_PACKET &rq->__lock irq_context: 0 sk_lock-AF_PACKET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET key irq_context: 0 sk_lock-AF_PACKET pcpu_lock irq_context: 0 sk_lock-AF_PACKET percpu_counters_lock irq_context: 0 sk_lock-AF_PACKET pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock ptype_lock irq_context: 0 slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 &net->packet.sklist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock ptype_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 fanout_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &x->wait#2 irq_context: 0 sk_lock-AF_PACKET &mm->mmap_lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim irq_context: 0 sk_lock-AF_PACKET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET vmap_area_lock irq_context: 0 sk_lock-AF_PACKET &____s->seqcount irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock purge_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &fp->aux->used_maps_mutex irq_context: 0 rlock-AF_PACKET irq_context: 0 wlock-AF_PACKET irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq &(&ifa->dad_work)->timer irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 &u->iolock &meta->lock irq_context: 0 &u->iolock kfence_freelist_lock irq_context: 0 sk_lock-AF_PACKET &c->lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock &c->lock irq_context: softirq &tx->clean_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock kfence_freelist_lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 quarantine_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx rcu_read_lock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock key irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq &tx->clean_lock &meta->lock irq_context: softirq &tx->clean_lock kfence_freelist_lock irq_context: softirq drivers/base/dd.c:321 irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->list_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->k_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_mutex irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_work irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &x->wait#10 irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#40 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 batched_entropy_u8.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 key#12 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->i_private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &nvmeq->sq_lock irq_context: hardirq &ei->i_completed_io_lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq rcu_read_lock &memcg->move_lock irq_context: hardirq rcu_read_lock &xa->xa_lock#9 irq_context: hardirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: hardirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: hardirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: hardirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: hardirq rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ei->i_completed_io_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &journal->j_wait_reserved irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ext4__ioend_wq[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ret->b_uptodate_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &base->lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &base->lock &obj_hash[i].lock irq_context: softirq &(&wb->bw_dwork)->timer irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) &wb->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) irq_context: softirq (&lapb->t1timer) &lapb->lock irq_context: softirq (&lapb->t1timer) &lapb->lock batched_entropy_u8.lock irq_context: softirq (&lapb->t1timer) &lapb->lock kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: softirq (&dev->watchdog_timer) irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_PACKET &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &meta->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem kfence_freelist_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &dir->lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &dentry->d_lock irq_context: 0 hostname_poll.wait.lock irq_context: softirq (&dom->period_timer) irq_context: softirq (&dom->period_timer) key#13 irq_context: softirq (&dom->period_timer) &p->sequence irq_context: softirq (&dom->period_timer) &obj_hash[i].lock irq_context: softirq (&dom->period_timer) &base->lock irq_context: softirq (&dom->period_timer) &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock once_lock irq_context: softirq rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock &____s->seqcount#8 irq_context: softirq rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock lock#8 irq_context: softirq rcu_read_lock rcu_read_lock id_table_lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &____s->seqcount#8 irq_context: softirq rcu_read_lock nl_table_lock irq_context: softirq rcu_read_lock rlock-AF_NETLINK irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 batched_entropy_u16.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 (&req->rsk_timer) irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &icsk->icsk_accept_queue.rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET &icsk->icsk_accept_queue.rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET clock-AF_INET irq_context: 0 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &base->lock irq_context: 0 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET fs_reclaim irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET &c->lock irq_context: 0 sk_lock-AF_INET pool_lock#2 irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock irq_context: 0 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: 0 crngs.lock base_crng.lock irq_context: 0 sk_lock-AF_INET &sd->defer_lock irq_context: softirq &sd->defer_lock irq_context: softirq (&icsk->icsk_delack_timer) irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 batched_entropy_u8.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 kfence_freelist_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 &u->iolock &dir->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 sk_lock-AF_INET remove_cache_srcu irq_context: 0 sk_lock-AF_INET remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET clock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock tcp_metrics_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 batched_entropy_u16.lock crngs.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->wr_wait irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &pipe->mutex/1 &obj_hash[i].lock irq_context: 0 &pipe->wr_wait &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_callback uidhash_lock irq_context: softirq rcu_callback percpu_counters_lock irq_context: softirq rcu_callback ucounts_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock irq_context: 0 &vma->vm_lock->lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->i_private_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock irq_context: 0 &vma->vm_lock->lock kfence_freelist_lock irq_context: softirq rcu_callback &zone->lock irq_context: softirq rcu_callback &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock key#12 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &journal->j_list_lock pool_lock#2 irq_context: 0 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &n->list_lock irq_context: 0 sk_lock-AF_INET &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 &pipe->wr_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 rcu_read_lock &stopper->lock irq_context: 0 rcu_read_lock &stop_pi_lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rq->__lock rcu_read_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &rq->__lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &____s->seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &n->list_lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stopper->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 tomoyo_ss irq_context: 0 sb_writers#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 sk_lock-AF_INET quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_list_lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 &c->lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 lock#5 irq_context: 0 rcu_read_lock &base->lock irq_context: 0 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &ret->b_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle key#3 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle key#4 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &sbi->s_error_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 kn->active#42 fs_reclaim irq_context: 0 kn->active#42 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#42 &c->lock irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock irq_context: 0 &futex_queues[i].lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 rcu_read_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &ep->lock &ep->wq irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &lock->wait_lock irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &p->lock irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 batched_entropy_u32.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &ep->mtx sysctl_lock irq_context: 0 &f->f_pos_lock sysctl_lock irq_context: 0 &f->f_pos_lock fs_reclaim irq_context: 0 &f->f_pos_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &zone->lock irq_context: 0 &f->f_pos_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 &ep->mtx kn->active#5 fs_reclaim irq_context: 0 &ep->mtx kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ep->mtx kn->active#5 pool_lock#2 irq_context: 0 &ep->mtx kn->active#5 &on->poll irq_context: 0 &ep->mtx kn->active#5 &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &of->mutex irq_context: 0 &f->f_pos_lock &p->lock &of->mutex kn->active#5 &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &of->mutex kn->active#5 param_lock irq_context: 0 &ep->mtx rcu_read_lock &on->poll irq_context: 0 &f->f_pos_lock &p->lock &c->lock irq_context: 0 &f->f_pos_lock &p->lock &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &n->list_lock irq_context: 0 &f->f_pos_lock &p->lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &p->lock pool_lock#2 irq_context: 0 &f->f_pos_lock &p->lock module_mutex irq_context: 0 sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET once_mutex irq_context: 0 sk_lock-AF_INET once_mutex crngs.lock irq_context: 0 sk_lock-AF_INET once_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET batched_entropy_u16.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&wq_watchdog_timer) &obj_hash[i].lock irq_context: softirq (&wq_watchdog_timer) &base->lock irq_context: softirq (&wq_watchdog_timer) &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ep->mtx &pipe->wr_wait irq_context: 0 rcu_read_lock tasklist_lock irq_context: 0 &ep->mtx rcu_read_lock &pipe->wr_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 sb_writers#4 &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#4 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &c->lock irq_context: 0 &type->s_umount_key#41/1 irq_context: 0 &type->s_umount_key#41/1 fs_reclaim irq_context: 0 &type->s_umount_key#41/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 pool_lock#2 irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#41/1 shrinker_mutex irq_context: 0 &type->s_umount_key#41/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#41/1 sb_lock irq_context: 0 &type->s_umount_key#41/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->s_umount_key#41/1 crngs.lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#41/1 &dentry->d_lock irq_context: 0 sb_writers#8 irq_context: 0 sb_writers#8 mount_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq#2 irq_context: 0 kn->active#43 fs_reclaim irq_context: 0 kn->active#43 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#43 pool_lock#2 irq_context: 0 sb_writers#8 fs_reclaim irq_context: 0 sb_writers#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &mm->mmap_lock irq_context: 0 sb_writers#8 &of->mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock &c->lock irq_context: 0 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 cgroup_mutex css_set_lock cgroup_file_kn_lock irq_context: 0 &type->s_umount_key#42/1 irq_context: 0 &type->s_umount_key#42/1 fs_reclaim irq_context: 0 &type->s_umount_key#42/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#42/1 shrinker_mutex irq_context: 0 &type->s_umount_key#42/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#42/1 sb_lock irq_context: 0 &type->s_umount_key#42/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->s_umount_key#42/1 crngs.lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#42/1 &dentry->d_lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#43 irq_context: 0 &type->s_umount_key#43 &x->wait#23 irq_context: 0 &type->s_umount_key#43 shrinker_mutex irq_context: 0 &type->s_umount_key#43 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock pool_lock#2 irq_context: 0 &type->s_umount_key#43 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#43 rename_lock.seqcount irq_context: 0 &type->s_umount_key#43 &dentry->d_lock irq_context: 0 &type->s_umount_key#43 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#43 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#43 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#43 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#43 inode_hash_lock irq_context: 0 &type->s_umount_key#43 inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#43 pool_lock#2 irq_context: 0 &type->s_umount_key#43 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#43 &dentry->d_lock/1 irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cgroup_mutex &x->wait#2 irq_context: 0 cgroup_mutex &rq->__lock irq_context: 0 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback percpu_ref_switch_waitq.lock irq_context: softirq rcu_callback rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 cgroup_mutex &n->list_lock irq_context: 0 cgroup_mutex &n->list_lock &c->lock irq_context: 0 cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) percpu_ref_switch_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &c->lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) percpu_ref_switch_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &cgrp->pidlist_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (wq_completion)cgroup_pidlist_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (work_completion)(&cgrp->release_agent_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pcpu_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) krc.lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &x->wait#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex kfence_freelist_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &c->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 &type->s_umount_key#42/1 &c->lock irq_context: 0 sb_writers#9 irq_context: 0 sb_writers#9 mount_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 cgroup_mutex cpuset_mutex irq_context: 0 cgroup_mutex cpuset_mutex callback_lock irq_context: 0 &type->s_umount_key#42/1 &n->list_lock irq_context: 0 &type->s_umount_key#42/1 &n->list_lock &c->lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 cgroup_mutex &dom->lock irq_context: 0 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cgroup_mutex batched_entropy_u8.lock irq_context: 0 cgroup_mutex kfence_freelist_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) &p->pi_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex callback_lock irq_context: 0 rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &rq->__lock &base->lock irq_context: 0 rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &dom->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) krc.lock &base->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq#2 irq_context: 0 kn->active#44 fs_reclaim irq_context: 0 kn->active#44 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 fs_reclaim irq_context: 0 sb_writers#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &mm->mmap_lock irq_context: 0 sb_writers#9 &of->mutex irq_context: 0 sb_writers#9 &obj_hash[i].lock irq_context: 0 kn->active#45 fs_reclaim irq_context: 0 kn->active#45 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#45 &c->lock irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex kn->active#45 cpu_hotplug_lock irq_context: 0 sb_writers#9 &of->mutex kn->active#45 cpu_hotplug_lock cpuset_mutex irq_context: 0 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#10 irq_context: 0 sb_writers#10 &mm->mmap_lock irq_context: 0 sb_writers#10 &attr->mutex irq_context: 0 sb_writers#10 &attr->mutex &mm->mmap_lock irq_context: 0 &type->s_umount_key#44 irq_context: 0 &type->s_umount_key#44 sb_lock irq_context: 0 &type->s_umount_key#44 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#11 irq_context: 0 sb_writers#11 fs_reclaim irq_context: 0 sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 pool_lock#2 irq_context: 0 sb_writers#11 &mm->mmap_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rename_lock.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 fs_reclaim irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 pool_lock#2 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &dentry->d_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &dentry->d_lock &wq irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &c->lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &____s->seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &s->s_inode_list_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 tk_core.seq.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 init_binfmt_misc.entries_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &wpan_dev->association_lock irq_context: 0 rtnl_mutex dev_addr_sem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tn->lock irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx &sec->lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex dev_addr_sem pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem rlock-AF_NETLINK irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock irq_context: 0 rtnl_mutex dev_addr_sem &pn->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem input_pool.lock irq_context: 0 rtnl_mutex _xmit_IEEE802154 irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &xa->xa_lock#16 irq_context: 0 &sb->s_type->i_mutex_key#10 genl_sk_destructing_waitq.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->beacon_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->mgmt_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &wdev->pmsr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem reg_indoor_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#14 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->w) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) pool_lock#2 irq_context: 0 &ep->mtx remove_cache_srcu irq_context: 0 &ep->mtx remove_cache_srcu quarantine_lock irq_context: 0 &ep->mtx remove_cache_srcu &c->lock irq_context: 0 &ep->mtx remove_cache_srcu &n->list_lock irq_context: 0 &ep->mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ep->mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &ep->mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &wb->list_lock irq_context: 0 &sbi->s_writepages_rwsem irq_context: 0 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 jbd2_handle irq_context: 0 &journal->j_wait_commit irq_context: 0 &sb->s_type->i_mutex_key#8 irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 init_mm.page_table_lock irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_alloc_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_alloc_mutex pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 batched_entropy_u32.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &((cluster_info + ci)->lock)/1 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex percpu_ref_switch_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 swap_avail_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 proc_poll_wait.lock irq_context: 0 swap_slots_cache_enable_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up swap_slots_cache_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 swap_slots_cache_enable_mutex swap_lock irq_context: 0 &____s->seqcount#4 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &sighand->siglock &c->lock irq_context: 0 rcu_read_lock &sighand->siglock &____s->seqcount irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#5 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#5 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 &mm->mmap_lock sb_writers#5 &wb->list_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &newf->file_lock &newf->resize_wait irq_context: 0 &kcov->lock irq_context: 0 &mm->mmap_lock &kcov->lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#7 irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &kcov->lock kcov_remote_lock irq_context: 0 &kcov->lock kcov_remote_lock pool_lock#2 irq_context: 0 &kcov->lock kcov_remote_lock &c->lock irq_context: 0 pid_caches_mutex irq_context: 0 pid_caches_mutex slab_mutex irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pid_caches_mutex slab_mutex pool_lock#2 irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 pid_caches_mutex slab_mutex &k->list_lock irq_context: 0 pid_caches_mutex slab_mutex lock irq_context: 0 pid_caches_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pid_caches_mutex slab_mutex &c->lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->s_umount_key#45 irq_context: 0 &type->s_umount_key#45 &rq->__lock irq_context: 0 &type->s_umount_key#45 sb_lock irq_context: 0 &type->s_umount_key#45 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 irq_context: 0 &sb->s_type->i_mutex_key#17 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &rq->__lock irq_context: 0 bt_proto_lock &sk->sk_peer_lock irq_context: 0 bt_proto_lock hci_sk_list.lock irq_context: 0 misc_mtx &base->lock irq_context: 0 misc_mtx &base->lock &obj_hash[i].lock irq_context: 0 (work_completion)(&(&data->open_timeout)->work) irq_context: 0 &data->open_mutex irq_context: 0 &data->open_mutex fs_reclaim irq_context: 0 &data->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex pool_lock#2 irq_context: 0 &data->open_mutex &pcp->lock &zone->lock irq_context: 0 &data->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex &____s->seqcount irq_context: 0 &data->open_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex &x->wait#9 irq_context: 0 &data->open_mutex hci_index_ida.xa_lock irq_context: 0 &data->open_mutex pcpu_alloc_mutex irq_context: 0 &data->open_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex wq_pool_mutex irq_context: 0 &data->open_mutex wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex pin_fs_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &data->open_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex irq_context: 0 &data->open_mutex gdp_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex gdp_mutex pool_lock#2 irq_context: 0 &data->open_mutex gdp_mutex &rq->__lock irq_context: 0 &data->open_mutex gdp_mutex lock irq_context: 0 &data->open_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &data->open_mutex lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex bus_type_sem irq_context: 0 &data->open_mutex bus_type_sem &rq->__lock irq_context: 0 &data->open_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &c->lock irq_context: 0 &data->open_mutex &n->list_lock irq_context: 0 &data->open_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex &rq->__lock irq_context: 0 &data->open_mutex &dev->power.lock irq_context: 0 &data->open_mutex dpm_list_mtx irq_context: 0 &data->open_mutex dpm_list_mtx &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex subsys mutex#74 irq_context: 0 &data->open_mutex subsys mutex#74 &rq->__lock irq_context: 0 &data->open_mutex subsys mutex#74 &k->k_lock irq_context: 0 &data->open_mutex &dev->devres_lock irq_context: 0 &data->open_mutex triggers_list_lock irq_context: 0 &data->open_mutex triggers_list_lock &rq->__lock irq_context: 0 &data->open_mutex leds_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &k->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex bus_type_sem irq_context: 0 &data->open_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &data->open_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &data->open_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex leds_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex.wait_lock irq_context: 0 &data->open_mutex &p->pi_lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &rfkill->lock irq_context: 0 &data->open_mutex hci_dev_list_lock irq_context: 0 &data->open_mutex tk_core.seq.seqcount irq_context: 0 &data->open_mutex hci_sk_list.lock irq_context: 0 &data->open_mutex (pm_chain_head).rwsem irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#45 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &data->open_mutex &list->lock#7 irq_context: 0 &data->open_mutex &data->read_wait irq_context: 0 &list->lock#7 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &p->alloc_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 hci_dev_list_lock irq_context: 0 &data->read_wait irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &list->lock#6 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1882 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1882 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1882 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1882 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1882 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1882 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1882 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1878 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1878 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ns->work) irq_context: 0 (wq_completion)events (work_completion)(&ns->work) sysctl_lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ns->work) sysctl_lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ns->work) keyring_name_lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&ns->work) stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#144 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &p->pi_lock irq_context: 0 &pool->lock &x->wait#10 irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock irq_context: 0 &hdev->req_lock pool_lock#2 irq_context: 0 &hdev->req_lock &list->lock#5 irq_context: 0 &hdev->req_lock &list->lock#6 irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->req_wait_q irq_context: 0 &hdev->req_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &base->lock irq_context: 0 &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (&timer.timer) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1883 irq_context: 0 (wq_completion)events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1883 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1883 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#19 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1675 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#19 irq_context: 0 fs_reclaim stock_lock irq_context: 0 fs_reclaim pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1676 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1676 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1676 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1676 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1883 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#536 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy46 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1729 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1883 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1883 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1883 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1879 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy46 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy45 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy45 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy45 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#47 irq_context: 0 (wq_completion)wg-kex-wg0#47 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 hci_dev_list_lock irq_context: 0 namespace_sem mnt_id_ida.xa_lock pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &undo_list->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex &nr_netdev_addr_lock_key irq_context: 0 rtnl_mutex listen_lock irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nl_table_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET/1 irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem &n->list_lock irq_context: 0 pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &c->lock irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 bt_proto_lock &c->lock irq_context: 0 bt_proto_lock &____s->seqcount irq_context: 0 bt_proto_lock &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 syslog_lock &rq->__lock irq_context: 0 syslog_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem (&timer.timer) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 &p->pi_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&lapb->t1timer) &lapb->lock &pcp->lock &zone->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#108 irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#107 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#107 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &base->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 pernet_ops_rwsem devices_rwsem irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &x->wait#19 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex sched_map-wait-type-override &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hsr->list_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->tt.commit_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex ptype_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &entry->crc_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex key#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->forw_bcast_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#21 irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: 0 sb_writers#4 &____s->seqcount#10 irq_context: 0 sb_writers#4 &(&net->ipv4.ping_group_range.lock)->lock irq_context: 0 sb_writers#4 &(&net->ipv4.ping_group_range.lock)->lock &____s->seqcount#10 irq_context: 0 misc_mtx &dir->lock irq_context: 0 rtnl_mutex &r->consumer_lock irq_context: 0 rtnl_mutex &r->consumer_lock &r->producer_lock irq_context: 0 rtnl_mutex failover_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &mm->mmap_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex rcu_read_lock lock#8 irq_context: 0 rtnl_mutex rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock irq_context: 0 rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &dir->lock#2 irq_context: 0 rtnl_mutex &ndev->lock pcpu_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex wq_pool_mutex &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex (&timer.timer) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 rtnl_mutex &br->hash_lock irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->hash_lock &c->lock irq_context: 0 rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex nf_hook_mutex irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex nf_hook_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 rtnl_mutex nf_hook_mutex &c->lock irq_context: 0 rtnl_mutex nf_hook_mutex &____s->seqcount irq_context: 0 rtnl_mutex j1939_netdev_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy41 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &ndev->req_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1883 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1883 &rq->__lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1883 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock deferred_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#291 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#291 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#291 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#291 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#291 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#291 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#291 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events fqdir_free_work &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#22 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#22 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &rq->__lock irq_context: softirq &hub->irq_urb_lock irq_context: softirq rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &meta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) kfence_freelist_lock irq_context: 0 (wq_completion)bond0#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1906 irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &base->lock irq_context: 0 &sb->s_type->i_lock_key#27 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &base->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &base->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1906 &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1906 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#60 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)nfc8_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1552 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#71 irq_context: 0 &data->open_mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex devtree_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex device_state_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex device_state_lock kernfs_notify_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &bus->devnum_next_mutex irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1788 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1788 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1788 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1788 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1788 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1784 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1789 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1789 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1789 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1789 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1789 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1789 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1789 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1785 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1550 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1550 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1550 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1549 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1549 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1549 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1790 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1790 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1790 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &rq->__lock irq_context: softirq (&hcd->rh_timer) irq_context: softirq (&hcd->rh_timer) &dum_hcd->dum->lock irq_context: softirq &hub->irq_urb_lock hcd_root_hub_lock irq_context: softirq &hub->irq_urb_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &x->wait#19 irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1688 irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1684 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1684 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#7 irq_context: 0 (wq_completion)hci2#7 &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1688 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1688 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1688 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_internal remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock pool_lock#2 irq_context: 0 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock nl_table_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem device_state_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem device_state_lock kernfs_notify_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex (console_sem).lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dum_hcd->dum->lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: 0 sb_writers#4 remove_cache_srcu irq_context: 0 sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock deferred_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1888 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &br->multicast_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->rexmit_timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock irq_context: softirq (&mp->timer) irq_context: softirq (&mp->timer) &br->multicast_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&mp->timer) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1888 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1888 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&pmctx->ip6_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&pmctx->ip4_mc_router_timer) irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &pl->lock key#12 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#57 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#71 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#75 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1790 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1790 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1786 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1791 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1791 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1791 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1787 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1787 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1787 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#57 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1792 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1792 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1792 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1788 irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1788 &rq->__lock irq_context: 0 &type->s_umount_key#47 binderfs_minors_mutex &rq->__lock irq_context: 0 &type->s_umount_key#47 binderfs_minors_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#3 irq_context: 0 pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh key#19 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1788 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1551 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 listen_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#145 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#145 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#146 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#146 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#146 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1551 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#145 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#146 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &devlink_port->type_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1551 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1551 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1908 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1908 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1908 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1908 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1906 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1906 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1906 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1901 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1627 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1627 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#468 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#75 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#75 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1794 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1794 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1794 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#468 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1626 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1626 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#468 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#322 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1556 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1556 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#322 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#322 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#322 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#468 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#321 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#321 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#321 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#321 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1909 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1909 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1907 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1902 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1628 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1628 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1628 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1628 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1627 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1627 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#469 irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss mount_lock irq_context: 0 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1550 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1550 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1550 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 &hdev->req_lock &c->lock irq_context: 0 rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#57 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#57 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 sb_writers#3 sb_internal rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1729 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1729 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#57 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#4 &n->list_lock irq_context: 0 sb_writers#4 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex key irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 rtnl_mutex key#15 irq_context: 0 rtnl_mutex &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&bat_priv->nc.work)->timer irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu irq_context: 0 pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#17 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex kernfs_idr_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex noop_qdisc.q.lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wq->mutex irq_context: 0 rtnl_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex hrtimer_bases.lock irq_context: 0 rtnl_mutex hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock irq_context: 0 rtnl_mutex init_lock slab_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock slab_mutex pool_lock#2 irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex init_lock slab_mutex &k->list_lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex init_lock slab_mutex &c->lock irq_context: 0 rtnl_mutex init_lock slab_mutex &____s->seqcount irq_context: 0 rtnl_mutex init_lock fs_reclaim irq_context: 0 rtnl_mutex init_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock &zone->lock irq_context: 0 rtnl_mutex init_lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock &base->lock irq_context: 0 rtnl_mutex init_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock crngs.lock irq_context: 0 pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &x->wait#21 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &br->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &c->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &pn->hash_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1595 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1595 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1595 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1595 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1596 irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&slave->notify_work)->timer irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1585 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1585 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1586 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1586 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1586 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1586 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1586 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1586 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1587 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1587 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &base->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1437 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1587 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1587 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1587 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1587 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1588 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1589 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1589 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1589 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1590 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1589 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1590 &rq->__lock irq_context: 0 &xa->xa_lock#20 stock_lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1590 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1590 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1590 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex lweventlist_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex crngs.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 rtnl_mutex ptype_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex _xmit_NONE irq_context: 0 rtnl_mutex lock#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#11 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &c->lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &____s->seqcount irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &hsr->list_lock irq_context: 0 rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex pin_fs_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex gdp_mutex &c->lock irq_context: 0 rtnl_mutex gdp_mutex &n->list_lock irq_context: 0 rtnl_mutex gdp_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &k->k_lock irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount irq_context: softirq &(&bat_priv->mcast.work)->timer irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock key#15 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key pool_lock#2 irq_context: softirq (&app->join_timer) irq_context: softirq (&app->join_timer) &app->lock irq_context: softirq (&app->join_timer) &list->lock#10 irq_context: softirq (&app->join_timer) batched_entropy_u32.lock irq_context: softirq (&app->join_timer) &obj_hash[i].lock irq_context: softirq (&app->join_timer) &base->lock irq_context: softirq (&app->join_timer) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &base->lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#18 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/1 irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: hardirq log_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: softirq (&app->join_timer)#2 irq_context: softirq (&app->join_timer)#2 &app->lock#2 irq_context: softirq (&app->join_timer)#2 &list->lock#11 irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &base->lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond0#25 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xfrm.xfrm_state_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1651 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &dev->mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: 0 sb_writers#7 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock pool_lock#2 irq_context: 0 rtnl_mutex req_lock irq_context: 0 rtnl_mutex &x->wait#11 irq_context: 0 rtnl_mutex subsys mutex#75 irq_context: 0 rtnl_mutex subsys mutex#75 &k->k_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1930 irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1565 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1565 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1565 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1565 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1566 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1566 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1906 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1416 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1416 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1455 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1455 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex stack_depot_init_mutex stack_depot_init_mutex.wait_lock irq_context: 0 rtnl_mutex stack_depot_init_mutex rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock irq_context: 0 kn->active#46 fs_reclaim irq_context: 0 kn->active#46 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#46 &c->lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock irq_context: 0 kn->active#47 fs_reclaim irq_context: 0 kn->active#47 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock nsim_bus_dev_ids.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex device_links_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fwnode_link_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &dev->devres_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_maps_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock irq_context: 0 (wq_completion)wg-kex-wg1#71 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1563 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1412 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1412 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1412 irq_context: 0 tomoyo_ss rcu_read_lock &____s->seqcount#3 irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 stock_lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1562 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1562 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1563 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1413 irq_context: 0 kn->active#53 &____s->seqcount#2 irq_context: 0 kn->active#53 &____s->seqcount irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1563 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1563 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1563 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1564 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1564 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1414 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1414 irq_context: 0 sk_lock-AF_PACKET &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1414 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1415 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1416 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1417 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1566 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1564 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_event_queue_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1415 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1416 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock (&timer.timer) irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock irq_context: 0 rtnl_mutex _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1729 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &base->lock &obj_hash[i].lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mount_lock mount_lock.seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock pool_lock#2 irq_context: 0 rcu_read_lock &wqueue->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] key irq_context: 0 &kernfs_locks->open_file_mutex[count] pcpu_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1497 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1497 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1497 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1497 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1497 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1497 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1497 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1497 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1718 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1718 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1718 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1718 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1718 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1717 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1717 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1717 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#38 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#38 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#38 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#38 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#9 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#382 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#265 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#382 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#382 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#265 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#265 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#265 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#265 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#265 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#265 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#382 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#265 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#265 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#265 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#382 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#382 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#382 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#49 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#49 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#49 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#49 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#49 irq_context: 0 &hdev->req_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#383 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#383 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#383 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#383 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1719 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1719 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1719 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1719 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1719 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1718 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#266 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#266 &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#266 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#266 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#266 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#266 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#266 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#266 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1727 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1726 irq_context: 0 rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &xt[i].mutex &rq->__lock cpu_asid_lock irq_context: 0 &sighand->siglock stock_lock irq_context: 0 &sighand->siglock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &n->list_lock irq_context: 0 &sighand->siglock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem quarantine_lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock irq_context: 0 &dev_instance->mutex remove_cache_srcu irq_context: 0 &dev_instance->mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1498 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1498 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex deferred_probe_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex probe_waitqueue.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock subsys mutex#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1498 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1498 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1498 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1498 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock irq_context: softirq &(&hwstats->traffic_dw)->timer irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex mount_lock irq_context: 0 rtnl_mutex mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#78 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#78 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#78 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1836 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1836 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1836 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1836 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1836 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1836 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1836 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#437 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1837 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1837 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1837 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1837 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#298 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#298 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#298 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#298 irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#438 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#438 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#438 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#438 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1838 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1838 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1838 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) pool_lock#2 irq_context: 0 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 rtnl_mutex &devlink_port->type_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount irq_context: 0 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1917 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1915 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1591 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1591 &rq->__lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1592 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1592 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sb_writers#3 jbd2_handle key irq_context: 0 sb_writers#3 jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#28 irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2017 irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc6_nci_rx_wq#95 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex freezer_mutex.wait_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex.wait_lock irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 &rcu_state.gp_wq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1529 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2017 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2017 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy54 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock key irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->orig_work)->timer irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 (wq_completion)events_power_efficient (gc_work).work irq_context: 0 (wq_completion)events_power_efficient (gc_work).work tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) key#18 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work "ratelimiter_table_lock" irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1498 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1498 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#39 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#39 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#50 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#50 irq_context: 0 misc_mtx remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)dm_bufio_cache irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) dm_bufio_clients_lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#50 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#50 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &n->list_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &c->lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1499 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1499 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &table->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &peer->endpoint_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_node_0 irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->periodic_timer) irq_context: softirq (&app->periodic_timer) &app->lock#2 irq_context: softirq (&app->periodic_timer) &app->lock#2 &obj_hash[i].lock irq_context: softirq (&app->periodic_timer) &app->lock#2 &base->lock irq_context: softirq (&app->periodic_timer) &app->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &c->lock irq_context: 0 rtnl_mutex _xmit_SIT irq_context: 0 rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &br->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &base->lock irq_context: 0 rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &____s->seqcount irq_context: softirq (&brmctx->ip6_own_query.timer) irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&brmctx->ip4_own_query.timer) irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock crngs.lock irq_context: softirq (&in_dev->mr_ifc_timer) irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &ul->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#7 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#8 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock irq_context: softirq (&in_dev->mr_ifc_timer) &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&br->gc_work)->timer irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex _xmit_TUNNEL irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ptype_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &cfs_rq->removed.lock irq_context: 0 sb_internal batched_entropy_u8.lock irq_context: 0 sb_internal kfence_freelist_lock irq_context: 0 sb_internal &meta->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2017 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2013 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2008 irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1551 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1551 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1551 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1551 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1552 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&tn->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&c->work)->work) irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex _xmit_IPGRE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krdsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rtn->rds_tcp_accept_w) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &icsk->icsk_accept_queue.rskq_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rds_tcp_conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem loop_conns_lock irq_context: 0 kn->active#46 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)l2tp irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_rx_wq#24 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount#2 irq_context: softirq rcu_callback rlock-AF_RXRPC irq_context: 0 (wq_completion)nfc9_nci_rx_wq#24 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#40 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->peer_keepalive_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->service_conn_reap_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#10 irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock pool_lock#2 irq_context: 0 &xt[i].mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex netpoll_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pn->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex _xmit_TUNNEL6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: softirq rcu_callback batched_entropy_u8.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#141 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override rcu_node_0 irq_context: 0 &fsnotify_mark_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex _xmit_IPGRE &c->lock irq_context: 0 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex devnet_rename_sem &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&hdev->cmd_timer)->timer irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 smack_known_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &c->lock irq_context: 0 smack_known_lock mmu_notifier_invalidate_range_start irq_context: 0 smack_known_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &base->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 virtio_9p_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1936 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1904 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1642 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 smack_known_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1936 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[3] irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1899 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &x->wait#9 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 (wq_completion)wg-kex-wg0#69 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &base->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1802 irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 nf_sockopt_mutex rcu_node_0 irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &cfs_rq->removed.lock irq_context: 0 nf_sockopt_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 nf_sockopt_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bond->stats_lock/1 irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &ndev->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#47 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev_instance->mutex &rq->__lock irq_context: 0 &dev_instance->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock (&timer.timer) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2018 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#303 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#303 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#303 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1596 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1444 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1444 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1444 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq &list->lock#12 irq_context: softirq rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1459 irq_context: 0 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1460 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1460 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1460 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1499 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#28 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &____s->seqcount irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#469 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#469 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#469 irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy66 irq_context: 0 (wq_completion)wg-kex-wg2#58 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#469 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc4_nci_rx_wq#469 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex stack_depot_init_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1609 irq_context: 0 &data->open_mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1593 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: softirq rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock kfence_freelist_lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &ep->mtx key irq_context: 0 &ep->mtx pcpu_lock irq_context: 0 &ep->mtx percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1461 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1461 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1461 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1461 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1461 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1462 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1462 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1462 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1462 irq_context: 0 ebt_mutex stock_lock irq_context: 0 wq_pool_attach_mutex wq_pool_attach_mutex.wait_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1463 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1319 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1463 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1463 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1463 irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1464 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: softirq (&p->forward_delay_timer) irq_context: softirq (&p->forward_delay_timer) &br->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &c->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock pool_lock#2 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_wait.lock irq_context: softirq (&n->timer) irq_context: softirq (&n->timer) &n->lock irq_context: softirq (&n->timer) &n->lock &obj_hash[i].lock irq_context: softirq (&n->timer) &n->lock &base->lock irq_context: softirq (&n->timer) &n->lock &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1464 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1464 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1464 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1464 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1465 irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tn->lock &rq->__lock irq_context: 0 rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1465 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1466 irq_context: 0 misc_mtx fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1466 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1322 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 misc_mtx &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1467 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1467 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1468 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1471 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (debug_obj_work).work quarantine_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &p->pi_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_node_0 irq_context: 0 &xt[i].mutex stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock deferred_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock (console_sem).lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1930 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_node_0 irq_context: 0 cb_lock genl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 &xt[i].mutex pool_lock irq_context: 0 &xt[i].mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&pmctx->ip6_own_query.timer) irq_context: softirq (&pmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&pmctx->ip4_own_query.timer) irq_context: softirq (&pmctx->ip4_own_query.timer) &br->multicast_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy63 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy63 &rq->__lock irq_context: 0 &child->perf_event_mutex rcu_node_0 irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#10 mount_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1794 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1794 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#424 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1795 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1795 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1795 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#425 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#425 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#425 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1796 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1796 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1796 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock deferred_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &br->hash_lock irq_context: softirq rcu_read_lock &br->hash_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock nl_table_lock irq_context: softirq rcu_read_lock &br->hash_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock irq_context: softirq rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->multicast_lock &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: softirq rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock deferred_lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1796 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1796 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1555 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1556 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1555 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1798 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy32 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1798 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1798 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1798 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1798 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1794 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1557 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1556 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1556 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1799 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1799 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1800 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1800 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1800 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock &mm->page_table_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 irq_context: 0 sb_writers#3 &mm->mmap_lock &sb->s_type->i_lock_key irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1800 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1800 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1796 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle key irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1801 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1801 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1801 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1797 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1802 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1802 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1798 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1798 &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy42 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy42 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kn->active#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy41 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 kn->active#32 &____s->seqcount#2 irq_context: 0 kn->active#32 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#20 irq_context: 0 (wq_completion)events fqdir_free_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#39 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1803 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1803 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1803 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1803 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1803 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: softirq drivers/regulator/core.c:6335 irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (regulator_init_complete_work).work irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->list_lock irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1593 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1594 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1442 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1594 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1594 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &hsr->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pmc->lock irq_context: 0 rtnl_mutex key irq_context: 0 rtnl_mutex percpu_counters_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1535 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: softirq (&hsr->announce_timer) irq_context: softirq (&hsr->announce_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &nn->netlink_tap_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1535 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock key#15 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex bit_wait_table + i irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &__ctx->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: hardirq rcu_read_lock &xa->xa_lock#9 key#11 irq_context: hardirq rcu_read_lock &xa->xa_lock#9 key#12 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq rcu_read_lock &xa->xa_lock#9 key#13 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex j1939_netdev_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock key irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock &list->lock#13 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &list->lock#13 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &retval->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &__ctx->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock pool_lock#2 irq_context: hardirq &rq_wait->wait irq_context: hardirq &rq_wait->wait &p->pi_lock irq_context: hardirq &rq_wait->wait &p->pi_lock &rq->__lock irq_context: hardirq &rq_wait->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 &mm->mmap_lock pool_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: softirq (&cb->timer) &rq_wait->wait irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock &rq->__lock irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex key#19 irq_context: 0 rtnl_mutex &bat_priv->tt.commit_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#12 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#13 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1799 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &rq_wait->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &dir->lock irq_context: 0 rtnl_mutex k-slock-AF_INET/1 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex k-slock-AF_INET#2 irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock key irq_context: 0 rtnl_mutex rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &wg->device_update_lock &dir->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET/1 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET#2 irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &wg->socket_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &list->lock#14 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1444 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1445 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1596 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1596 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1596 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1596 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1597 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1597 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1445 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1445 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1445 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1446 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1446 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1597 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1597 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1597 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1598 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1598 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1446 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1446 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1446 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1446 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1447 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1598 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1598 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1598 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1598 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1599 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1599 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1447 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1447 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1447 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1447 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1592 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1593 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1593 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1593 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1593 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1593 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1593 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1441 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1441 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1441 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1441 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1442 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1442 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1594 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1594 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1442 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1442 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1442 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1443 irq_context: 0 &xa->xa_lock#20 &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1595 irq_context: softirq rcu_read_lock rcu_read_lock &r->producer_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1595 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1447 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1448 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1599 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1599 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1599 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1599 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1600 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1600 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1448 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1448 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1448 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1443 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1443 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1443 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1443 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1590 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1438 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1438 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1438 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1439 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1439 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xa->xa_lock#20 pool_lock#2 irq_context: 0 &xa->xa_lock#20 &obj_hash[i].lock irq_context: 0 &iopt->domains_rwsem irq_context: 0 &iopt->domains_rwsem &iopt->iova_rwsem irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1591 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1591 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1439 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1439 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1439 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1439 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1440 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1591 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1592 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1592 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1440 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1440 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1440 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1440 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1441 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1441 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1443 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1443 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1444 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1448 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1448 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1449 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1600 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1600 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1600 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1600 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1601 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1601 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1601 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1449 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1449 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1449 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1449 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1449 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1450 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1450 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock rcu_read_lock rcu_node_0 irq_context: 0 &hdev->req_lock rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (wq_completion)hci1#10 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq &keypair->receiving_counter.lock irq_context: softirq &peer->keypairs.keypair_update_lock irq_context: softirq &list->lock#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1601 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1601 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1602 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1602 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1450 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1450 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1450 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1451 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#7 irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 misc_mtx rfkill_global_mutex irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx rfkill_global_mutex pool_lock#2 irq_context: 0 misc_mtx rfkill_global_mutex &rfkill->lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: softirq rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 cb_lock genl_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex hwsim_radio_lock irq_context: 0 cb_lock genl_mutex &x->wait#9 irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex gdp_mutex irq_context: 0 cb_lock genl_mutex gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: softirq rcu_read_lock_bh &base->lock irq_context: softirq rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex subsys mutex#53 irq_context: 0 cb_lock genl_mutex subsys mutex#53 &k->k_lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex device_links_lock irq_context: 0 cb_lock genl_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex deferred_probe_mutex irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex wq_pool_mutex irq_context: 0 cb_lock genl_mutex wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex crngs.lock irq_context: 0 cb_lock genl_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#14 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh key#19 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 rtnl_mutex _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&tun->flow_gc_timer) irq_context: softirq (&tun->flow_gc_timer) &tun->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex.wait_lock irq_context: 0 cb_lock genl_mutex pin_fs_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock key#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: softirq (&app->join_timer) batched_entropy_u32.lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#4 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &retval->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx failover_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ndev->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem irq_context: 0 cb_lock genl_mutex inet6addr_chain.lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#3 &n->list_lock irq_context: 0 sb_writers#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&dwork->timer) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#9 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &dentry->d_lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex.wait_lock irq_context: 0 cb_lock &p->pi_lock irq_context: 0 cb_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override stock_lock irq_context: 0 fill_pool_map-wait-type-override key irq_context: 0 fill_pool_map-wait-type-override pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 fill_pool_map-wait-type-override percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &lo->lo_mutex &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &lo->lo_mutex &obj_hash[i].lock irq_context: 0 &disk->open_mutex &lo->lo_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fill_pool_map-wait-type-override pcpu_lock stock_lock irq_context: 0 &xt[i].mutex fs_reclaim pool_lock#2 irq_context: 0 &xt[i].mutex fs_reclaim stock_lock irq_context: 0 &xt[i].mutex fs_reclaim key irq_context: 0 &xt[i].mutex fs_reclaim pcpu_lock irq_context: 0 &xt[i].mutex fs_reclaim percpu_counters_lock irq_context: 0 &xt[i].mutex fs_reclaim pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rtnl_mutex.wait_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx key irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#15 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &list->lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex net_rwsem nl_table_lock irq_context: 0 rtnl_mutex net_rwsem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex net_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex net_rwsem &____s->seqcount irq_context: 0 rtnl_mutex net_rwsem nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &list->lock#2 irq_context: 0 &type->s_umount_key#46/1 irq_context: 0 &type->s_umount_key#46/1 fs_reclaim irq_context: 0 &type->s_umount_key#46/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &c->lock irq_context: 0 &type->s_umount_key#46/1 pool_lock#2 irq_context: 0 &type->s_umount_key#46/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#46/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#46/1 shrinker_mutex irq_context: 0 &type->s_umount_key#46/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#46/1 sb_lock irq_context: 0 &type->s_umount_key#46/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#46/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#46/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#46/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 binderfs_minors_mutex irq_context: 0 &type->s_umount_key#46/1 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &type->s_umount_key#46/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &____s->seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rename_lock.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 fs_reclaim irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 pool_lock#2 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &dentry->d_lock &wq irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock iunique_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_file_kn_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex task_group_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#6 irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq irq_context: 0 kn->active#48 fs_reclaim irq_context: 0 kn->active#48 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#49 fs_reclaim irq_context: 0 kn->active#49 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &x->wait#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_wait.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#15 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 pidmap_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET key irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#39 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pcpu_lock stock_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#27 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1966 irq_context: 0 lock prog_idr_lock irq_context: 0 lock prog_idr_lock pool_lock#2 irq_context: 0 bpf_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1966 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1966 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1966 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &k->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1962 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1804 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 nfc_devlist_mutex gdp_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1957 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1679 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1679 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1678 irq_context: 0 ebt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1549 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1775 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1775 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1775 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1775 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1775 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1678 irq_context: softirq rcu_callback rlock-AF_NETLINK irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1933 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1967 irq_context: 0 (wq_completion)bond0#19 irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#74 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1957 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim &rq->__lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock &rq->__lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (wq_completion)hci2#10 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: softirq &list->lock#16 irq_context: softirq rcu_read_lock lock#6 irq_context: softirq rcu_read_lock lock#6 kcov_remote_lock irq_context: softirq rcu_read_lock &ifibss->incomplete_lock irq_context: softirq rcu_read_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock irq_context: softirq rcu_read_lock &local->rx_path_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1933 &rq->__lock irq_context: 0 pernet_ops_rwsem &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: softirq &(&conn->disc_work)->timer irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#492 irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#19 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1660 irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_node_0 irq_context: 0 rcu_read_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_internal jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#47 irq_context: 0 (wq_completion)wg-kex-wg1#47 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#48 irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev_instance->mutex fs_reclaim &rq->__lock irq_context: 0 &dev_instance->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#15 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &xa->xa_lock#9 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &fsnotify_mark_srcu irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex &pool->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex patch_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex callback_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex percpu_counters_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#7 irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq irq_context: 0 kn->active#50 fs_reclaim irq_context: 0 kn->active#50 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#50 &c->lock irq_context: 0 kn->active#50 &____s->seqcount#2 irq_context: 0 kn->active#50 &____s->seqcount irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex cpuset_attach_wq.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock pool_lock#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &pool->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1804 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1804 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount#2 irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &xa->xa_lock#5 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &xa->xa_lock#5 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#5 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#5 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#51 fs_reclaim irq_context: 0 kn->active#51 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#51 stock_lock irq_context: 0 kn->active#51 pool_lock#2 irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#51 &c->lock irq_context: 0 kn->active#52 fs_reclaim irq_context: 0 kn->active#52 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#52 stock_lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex kn->active#52 memcg_max_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &xa->xa_lock#5 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &xa->xa_lock#5 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex devcgroup_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: softirq rcu_read_lock hwsim_radio_lock irq_context: softirq rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &list->lock#16 irq_context: softirq rcu_read_lock &local->rx_path_lock &list->lock#15 irq_context: softirq rcu_read_lock &local->rx_path_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 kcov_remote_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock irq_context: 0 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#5 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#5 &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#5 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 kn->active#50 stock_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock &newf->file_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock pool_lock irq_context: 0 &xa->xa_lock#5 pool_lock#2 irq_context: 0 nf_hook_mutex irq_context: 0 nf_hook_mutex fs_reclaim irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_hook_mutex stock_lock irq_context: 0 nf_hook_mutex pool_lock#2 irq_context: 0 nf_hook_mutex &c->lock irq_context: 0 nf_hook_mutex &____s->seqcount#2 irq_context: 0 nf_hook_mutex &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 ebt_mutex &mm->mmap_lock fs_reclaim irq_context: 0 ebt_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 ebt_mutex &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock stock_lock irq_context: 0 ebt_mutex &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &xt[i].mutex vmap_area_lock irq_context: 0 &xt[i].mutex &per_cpu(xt_recseq, i) irq_context: 0 &xt[i].mutex &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock irq_context: 0 &vma->vm_lock->lock stock_lock irq_context: 0 nf_nat_proto_mutex irq_context: 0 nf_nat_proto_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex &c->lock irq_context: 0 nf_nat_proto_mutex &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex &____s->seqcount irq_context: 0 nf_nat_proto_mutex nf_hook_mutex irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex nf_hook_mutex stock_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex cpu_hotplug_lock irq_context: 0 nf_nat_proto_mutex &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 nf_nat_proto_mutex stock_lock irq_context: 0 pcpu_alloc_mutex fs_reclaim irq_context: 0 pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pcpu_alloc_mutex &____s->seqcount irq_context: 0 pcpu_alloc_mutex free_vmap_area_lock irq_context: 0 pcpu_alloc_mutex vmap_area_lock irq_context: 0 pcpu_alloc_mutex &c->lock irq_context: 0 pcpu_alloc_mutex &____s->seqcount#2 irq_context: 0 pcpu_alloc_mutex init_mm.page_table_lock irq_context: 0 pcpu_alloc_mutex &pcp->lock &zone->lock irq_context: 0 pcpu_alloc_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 stock_lock irq_context: 0 &ep->mtx rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#9 irq_context: 0 &vma->vm_lock->lock &sb->s_type->i_lock_key irq_context: 0 &vma->vm_lock->lock &info->lock irq_context: 0 &vma->vm_lock->lock lock#4 irq_context: 0 &vma->vm_lock->lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock tk_core.seq.seqcount irq_context: 0 &vma->vm_lock->lock mount_lock irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#5 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#5 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#5 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount#2 irq_context: 0 loop_validate_mutex irq_context: 0 loop_validate_mutex &lo->lo_mutex irq_context: 0 &fsnotify_mark_srcu fs_reclaim irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem stock_lock irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 &mm->mmap_lock stock_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#5 irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#5 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 stock_lock irq_context: softirq rcu_callback stock_lock irq_context: 0 &r->consumer_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock stock_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 &group->notification_waitq &ep->lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 misc_mtx nfc_index_ida.xa_lock irq_context: 0 misc_mtx &____s->seqcount#2 irq_context: 0 misc_mtx pcpu_alloc_mutex irq_context: 0 misc_mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 misc_mtx cpu_hotplug_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 misc_mtx kthread_create_lock irq_context: 0 misc_mtx &x->wait irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &____s->seqcount irq_context: 0 misc_mtx hrtimer_bases.lock irq_context: 0 misc_mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 misc_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&hsr->prune_timer) irq_context: softirq (&hsr->prune_timer) &hsr->list_lock irq_context: softirq (&hsr->prune_timer) &obj_hash[i].lock irq_context: softirq (&hsr->prune_timer) &base->lock irq_context: softirq (&hsr->prune_timer) &base->lock &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 misc_mtx wq_pool_mutex irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 kn->active#4 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &c->lock irq_context: 0 cb_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex net_rwsem quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &lock->wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx nfc_devlist_mutex pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex &k->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &k->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem irq_context: 0 misc_mtx nfc_devlist_mutex sysfs_symlink_target_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex &dev->power.lock irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx irq_context: 0 misc_mtx nfc_devlist_mutex &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &k->k_lock irq_context: 0 misc_mtx &k->k_lock irq_context: 0 misc_mtx llcp_devices_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex &x->wait#9 irq_context: 0 misc_mtx &dev->mutex &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &k->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &dev->power.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rfkill->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 misc_mtx &n->list_lock irq_context: 0 misc_mtx &n->list_lock &c->lock irq_context: 0 misc_mtx nl_table_lock irq_context: 0 misc_mtx nl_table_wait.lock irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount#2 irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#9 stock_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &____s->seqcount#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &lock->wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock stock_lock irq_context: 0 &ndev->req_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &ndev->req_lock &wq->mutex irq_context: 0 &ndev->req_lock &wq->mutex &pool->lock irq_context: 0 &ndev->req_lock &wq->mutex &x->wait#10 irq_context: 0 &ndev->req_lock (&ndev->cmd_timer) irq_context: 0 &ndev->req_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock &base->lock irq_context: 0 &ndev->req_lock (&ndev->data_timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &c->lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex kernfs_idr_lock irq_context: 0 &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex subsys mutex#40 irq_context: 0 &dev->mutex subsys mutex#40 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &rfkill->lock irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex triggers_list_lock irq_context: 0 &dev->mutex leds_list_lock irq_context: 0 &local->sockets.lock irq_context: 0 &local->raw_sockets.lock irq_context: 0 (&local->link_timer) irq_context: 0 (work_completion)(&local->tx_work) irq_context: 0 (work_completion)(&local->rx_work) irq_context: 0 (work_completion)(&local->timeout_work) irq_context: 0 (&local->sdreq_timer) irq_context: 0 (work_completion)(&local->sdreq_timeout_work) irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 nfc_devlist_mutex kernfs_idr_lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &k->k_lock klist_remove_lock irq_context: 0 nfc_devlist_mutex &x->wait#9 irq_context: 0 nfc_devlist_mutex deferred_probe_mutex irq_context: 0 nfc_devlist_mutex device_links_lock irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 nfc_devlist_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1499 irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &dev->mutex &sem->wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 pcpu_lock stock_lock irq_context: 0 tasklist_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#3 &s->s_inode_list_lock irq_context: 0 sb_writers#3 sb_internal irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal pool_lock#2 irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 sb_internal &obj_hash[i].lock irq_context: 0 sb_writers#3 inode_hash_lock irq_context: 0 sb_writers#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &fsnotify_mark_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &wb->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &obj_hash[i].lock irq_context: 0 &xt[i].mutex init_mm.page_table_lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &xt[i].mutex &n->list_lock irq_context: 0 &xt[i].mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 &c->lock irq_context: 0 sb_writers#7 kn->active#4 &rq->__lock irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock irq_context: softirq rcu_callback put_task_map-wait-type-override per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &xattrs->lock irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock rcu_node_0 irq_context: softirq slock-AF_INET#2 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &____s->seqcount#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem quarantine_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem pool_lock#2 irq_context: 0 kn->active#49 &c->lock irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 prog_idr_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim key irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 &base->lock irq_context: softirq slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#9 &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss &____s->seqcount#2 irq_context: 0 &xt[i].mutex &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &p->lock &of->mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &x->wait#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2018 &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &wq->mutex &rq->__lock irq_context: 0 &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#24 &lru->node[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 nf_nat_proto_mutex &rq->__lock irq_context: 0 nf_nat_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2018 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1609 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1610 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1456 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1456 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1456 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &sem->wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mdev->req_queue_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE slock-AF_ROSE irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1930 irq_context: 0 &dev->mutex dpm_list_mtx &rq->__lock irq_context: 0 &dev->mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &base->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &n->list_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: softirq (&app->join_timer) batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1651 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1933 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1933 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1931 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1930 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1931 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1931 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1926 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1926 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1926 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#329 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#329 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#329 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1650 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1650 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1650 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#329 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#328 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#35 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1660 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1660 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1660 irq_context: softirq rcu_read_lock &br->hash_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock &br->hash_lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1457 irq_context: 0 sb_writers#9 &n->list_lock irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1924 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1924 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1924 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1927 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1923 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1610 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1457 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1457 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1457 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &meta->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle key irq_context: 0 sb_writers#3 sb_internal jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1457 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1458 irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1771 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1610 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1610 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1611 irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 irq_context: 0 rtnl_mutex team->team_lock_key#36 fs_reclaim irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1459 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#440 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#440 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#440 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#440 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#440 &rq->__lock irq_context: 0 put_task_map-wait-type-override#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#440 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#440 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1840 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1840 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1840 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1840 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1840 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1840 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1840 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1836 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#441 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#441 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#301 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#301 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#301 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#301 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#301 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#301 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#441 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#441 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#441 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#441 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#441 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#441 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#80 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#9 &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1967 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex devnet_rename_sem batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &pcp->lock &zone->lock irq_context: 0 vsock_table_lock irq_context: 0 sk_lock-AF_VSOCK irq_context: 0 sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 sk_lock-AF_VSOCK &mm->mmap_lock irq_context: 0 slock-AF_VSOCK irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock irq_context: 0 sb_writers#3 fs_reclaim irq_context: 0 sb_writers#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_lock_key#20 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK vsock_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK clock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_VSOCK irq_context: 0 tasklist_lock &p->alloc_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1535 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1536 irq_context: 0 &hdev->req_lock (wq_completion)hci2#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1536 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1536 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1388 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex quarantine_lock irq_context: 0 &sig->cred_guard_mutex &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1537 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1537 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1537 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1538 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1538 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1538 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1538 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1539 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1539 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1539 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1539 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1390 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1391 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1539 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1539 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1540 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1540 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &meta->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#328 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1540 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1391 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1391 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem rcu_node_0 irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu pool_lock#2 irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1391 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_prealloc_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1540 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1540 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1544 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy44 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1552 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock cpu_asid_lock irq_context: 0 put_task_map-wait-type-override#2 &pcp->lock &zone->lock irq_context: 0 put_task_map-wait-type-override#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 put_task_map-wait-type-override#2 &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (work_completion)(&local->rx_work) &rq->__lock irq_context: 0 (work_completion)(&local->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock rcu_read_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_node_0 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1540 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1540 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1542 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1403 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1403 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1403 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#13 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 kn->active#14 &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1546 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1546 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1546 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx misc_mtx.wait_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1547 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1548 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1548 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &n->list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &____s->seqcount#2 irq_context: 0 rtnl_mutex &xa->xa_lock#4 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem kfence_freelist_lock irq_context: 0 rtnl_mutex devnet_rename_sem &meta->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#328 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#328 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1650 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1649 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1533 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1533 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1547 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 tomoyo_ss rcu_node_0 irq_context: 0 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1649 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1649 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#97 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#97 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1649 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1649 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1649 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#97 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#96 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &pcp->lock &zone->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 cb_lock remove_cache_srcu &c->lock irq_context: 0 cb_lock remove_cache_srcu &n->list_lock irq_context: 0 cb_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#481 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#481 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#481 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 udc_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 &hashinfo->lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#80 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#80 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#80 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1841 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1841 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1841 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1841 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1841 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1841 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1841 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1837 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1585 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1585 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1584 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1584 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1584 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1584 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1586 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1586 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1585 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1585 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#442 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#442 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#442 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#442 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1587 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1587 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1587 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1586 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1586 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1586 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1586 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#302 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#302 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#302 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#302 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#81 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#81 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#81 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#302 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#81 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1842 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1842 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1842 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1838 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#302 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#303 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#303 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#303 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#303 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1843 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1843 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#303 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1843 irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1843 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1843 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1661 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1661 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1839 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1839 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1839 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1588 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1588 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1638 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1588 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1638 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1588 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#478 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1588 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1587 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1587 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#55 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2001 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1587 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#443 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#443 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#443 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#443 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#443 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#443 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#304 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#304 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#304 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1589 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1589 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1588 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1588 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1590 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1590 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1589 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1589 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1589 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1589 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1589 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#305 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#305 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#305 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#305 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#304 irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#304 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#304 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#444 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#444 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1996 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#63 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#64 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#32 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 jbd2_handle &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#56 irq_context: 0 misc_mtx (wq_completion)nfc6_nci_rx_wq#95 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc6_nci_rx_wq#95 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#36 netpoll_srcu irq_context: 0 misc_mtx.wait_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#36 net_rwsem irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1917 irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 cb_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock rcu_read_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1460 irq_context: 0 (wq_completion)events (work_completion)(&w->w) &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#36 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#36 &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx rfkill_global_mutex &c->lock irq_context: 0 cb_lock genl_mutex genl_mutex.wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_internal remove_cache_srcu rcu_node_0 irq_context: 0 sb_internal remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_internal remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#300 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#300 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#300 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#300 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#300 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#300 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#300 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#300 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1839 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1839 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1839 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1839 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1839 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1835 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#439 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#439 irq_context: softirq &(&conn->info_timer)->timer irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#439 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#439 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1584 irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) &conn->chan_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1584 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1583 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1583 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#36 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#36 &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 udc_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#481 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#481 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#481 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1614 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1614 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1614 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu &base->lock irq_context: 0 tomoyo_ss remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (debug_obj_work).work &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1462 irq_context: 0 (wq_completion)events (debug_obj_work).work kfence_freelist_lock irq_context: 0 tomoyo_ss remove_cache_srcu &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1462 &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1615 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1615 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1615 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy48 irq_context: 0 rtnl_mutex _xmit_IPGRE &n->list_lock irq_context: 0 rtnl_mutex _xmit_IPGRE &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1616 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1616 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1617 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1617 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1464 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#444 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#444 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#444 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1844 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1844 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1844 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1840 irq_context: 0 (wq_completion)events fqdir_free_work quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#82 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#82 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#82 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#82 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#58 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#58 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#58 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#58 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#58 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#83 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#83 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#83 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#83 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1464 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy47 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1617 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#83 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#83 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1617 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1617 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy47 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1465 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1618 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1845 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1845 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1845 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#113 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#113 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1661 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 &p->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &xt[i].mutex rcu_read_lock rcu_node_0 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &ep->mtx &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &c->lock irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#6 &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount irq_context: 0 sb_writers#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &____s->seqcount#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &____s->seqcount irq_context: 0 kn->active#48 &c->lock irq_context: 0 kn->active#48 &n->list_lock irq_context: 0 kn->active#48 &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &dev_instance->mutex &____s->seqcount#2 irq_context: 0 &dev_instance->mutex &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock irq_context: 0 &dev->mutex &____s->seqcount#2 irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &____s->seqcount#2 irq_context: 0 nfc_devlist_mutex &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount#2 irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1618 &rq->__lock irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock irq_context: softirq (&dum_hcd->timer) &dev->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem nl_table_wait.lock irq_context: 0 kn->active#52 &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 &type->s_umount_key#46/1 &n->list_lock irq_context: 0 &type->s_umount_key#46/1 &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#46/1 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#46/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#46/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &c->lock irq_context: 0 ebt_mutex &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &____s->seqcount irq_context: 0 &xt[i].mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &n->list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &n->list_lock &c->lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &dev->tx_global_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &sch->q.lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#48 &____s->seqcount#2 irq_context: 0 kn->active#48 &____s->seqcount irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#50 &n->list_lock irq_context: 0 kn->active#50 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &____s->seqcount irq_context: 0 misc_mtx &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: softirq (&peer->timer_persistent_keepalive) irq_context: softirq (&peer->timer_persistent_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) &list->lock#14 irq_context: softirq (&peer->timer_persistent_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &xt[i].mutex remove_cache_srcu irq_context: 0 &xt[i].mutex remove_cache_srcu quarantine_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &c->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &n->list_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: softirq &tx->clean_lock quarantine_lock irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock irq_context: 0 rcu_read_lock &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &c->lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &knet->mutex irq_context: 0 &mux->lock irq_context: 0 &mux->rx_lock irq_context: 0 sync_timeline_list_lock irq_context: 0 &obj->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 driver_id_numbers.xa_lock irq_context: 0 &dev->lock irq_context: 0 uevent_sock_mutex fs_reclaim irq_context: 0 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 uevent_sock_mutex nl_table_lock irq_context: 0 uevent_sock_mutex &c->lock irq_context: 0 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 uevent_sock_mutex &____s->seqcount irq_context: 0 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock irq_context: 0 udc_lock irq_context: 0 udc_lock (console_sem).lock irq_context: 0 udc_lock console_lock console_srcu console_owner_lock irq_context: 0 udc_lock console_lock console_srcu console_owner irq_context: 0 udc_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 udc_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 udc_lock &rq->__lock irq_context: 0 udc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu irq_context: 0 misc_mtx remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx remove_cache_srcu &c->lock irq_context: 0 misc_mtx remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &cfs_rq->removed.lock irq_context: 0 misc_mtx batched_entropy_u8.lock irq_context: 0 misc_mtx kfence_freelist_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &dev->lock (console_sem).lock irq_context: 0 &dev->lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->lock console_lock console_srcu console_owner irq_context: 0 &dev->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &dev->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &bgl->locks[i].lock irq_context: 0 &fc->uapi_mutex irq_context: 0 &fc->uapi_mutex fs_reclaim irq_context: 0 &fc->uapi_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex pool_lock#2 irq_context: 0 &info->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount irq_context: 0 &sb->s_type->i_lock_key#16 irq_context: 0 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM clock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&kcm->tx_work) irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock rlock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &knet->mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET slock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &pnsocks.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET resource_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &x->wait#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &pcp->lock &zone->lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu pool_lock#2 irq_context: 0 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_PHONET irq_context: 0 &obj->lock tk_core.seq.seqcount irq_context: 0 &resv_map->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &sem->wait_lock irq_context: 0 nfc_devlist_mutex &p->pi_lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 &dev_instance->mutex &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &____s->seqcount#2 irq_context: 0 sb_writers#7 kn->active#4 &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 pool_lock#2 irq_context: 0 sb_writers#7 iattr_mutex &rq->__lock irq_context: 0 sb_writers#7 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &____s->seqcount#2 irq_context: 0 sb_writers#7 &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle key#4 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &sk->sk_lock.wq irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex quarantine_lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 pool_lock#2 irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 nfc_devlist_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex.wait_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) pool_lock#2 irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (work_completion)(&data->suspend_work) &rq->__lock irq_context: 0 (work_completion)(&data->suspend_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 sb_writers#4 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dom->period_timer) &p->sequence key#13 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &group->mark_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tun->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx batched_entropy_u8.lock crngs.lock irq_context: 0 put_task_map-wait-type-override#2 &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &u->iolock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 quarantine_lock irq_context: 0 &dev_instance->mutex &n->list_lock irq_context: 0 &dev_instance->mutex &n->list_lock &c->lock irq_context: 0 kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sdp->delay_work) irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events &rq->__lock irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events fqdir_free_work pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: softirq (&ndev->rs_timer) irq_context: softirq (&ndev->rs_timer) &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) pool_lock#2 irq_context: softirq (&ndev->rs_timer) &dir->lock#2 irq_context: softirq (&ndev->rs_timer) &ul->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#11 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock cpu_asid_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (work_completion)(&udc->vbus_work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#7 kn->active#4 &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &bgl->locks[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex quarantine_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &pcp->lock &zone->lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &xa->xa_lock#5 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &xa->xa_lock#5 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &xa->xa_lock#5 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ret->b_state_lock &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 kn->active#14 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex.wait_lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex.wait_lock irq_context: 0 misc_mtx wq_pool_mutex.wait_lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex &n->list_lock irq_context: 0 &dev->mutex &n->list_lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 rcu_read_lock stock_lock irq_context: 0 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#481 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock nsim_bus_dev_ids.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->list_lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &rq->__lock irq_context: 0 nf_sockopt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &u->iolock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 udc_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#481 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1934 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &c->lock irq_context: 0 ebt_mutex &rq->__lock irq_context: 0 ebt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &lock->wait_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &obj_hash[i].lock pool_lock irq_context: 0 key#21 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1404 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->tt.work)->timer irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1549 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#114 irq_context: 0 &tun->lock irq_context: 0 &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#15 irq_context: softirq rcu_callback put_task_map-wait-type-override &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.req_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.roam_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock quarantine_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 sb_internal &rq->__lock irq_context: 0 sb_writers#3 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2018 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2014 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount#2 irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2014 &rq->__lock irq_context: 0 sb_writers#3 sb_internal &n->list_lock irq_context: 0 sb_writers#3 sb_internal &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &rq->__lock irq_context: 0 file_rwsem &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#542 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 &dev->mutex device_links_lock &rq->__lock irq_context: 0 &dev->mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 &dev->mutex device_links_lock device_links_lock.wait_lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex device_links_lock.wait_lock irq_context: 0 nfc_devlist_mutex device_links_lock.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_base_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &____s->seqcount#2 irq_context: softirq rcu_read_lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kcov->lock kcov_remote_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex sched_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock rcu_read_lock &sighand->siglock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 tomoyo_ss tomoyo_policy_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->lock irq_context: 0 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_internal irq_context: 0 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->i_private_lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_internal &____s->seqcount irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 &type->s_umount_key#47 irq_context: 0 &type->s_umount_key#47 &x->wait#23 irq_context: 0 &type->s_umount_key#47 shrinker_mutex irq_context: 0 &type->s_umount_key#47 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#47 pool_lock#2 irq_context: 0 &type->s_umount_key#47 rename_lock.seqcount irq_context: 0 &type->s_umount_key#47 &dentry->d_lock irq_context: 0 &type->s_umount_key#47 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#47 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#47 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#47 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#47 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#47 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#47 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#47 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#47 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#47 &rq->__lock irq_context: 0 &type->s_umount_key#47 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock irq_context: 0 &type->s_umount_key#47 binderfs_minors_mutex irq_context: 0 &type->s_umount_key#47 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &mq_lock irq_context: 0 (wq_completion)events free_ipc_work irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &x->wait#2 irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kcov_remote_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work mount_lock irq_context: 0 (wq_completion)events free_ipc_work mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &x->wait#23 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 shrinker_mutex irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 rename_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &sb->s_type->i_lock_key#19 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &s->s_inode_list_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &xa->xa_lock#9 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#48 &dentry->d_lock/1 irq_context: 0 (wq_completion)events free_ipc_work sb_lock irq_context: 0 (wq_completion)events free_ipc_work unnamed_dev_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work list_lrus_mutex irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#5 irq_context: 0 (wq_completion)events free_ipc_work pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sb_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sb_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work mnt_id_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work &ids->rwsem irq_context: 0 (wq_completion)events free_ipc_work (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work percpu_counters_lock irq_context: 0 (wq_completion)events free_ipc_work pcpu_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock sysctl_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)events free_ipc_work &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)events free_ipc_work &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work rename_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &s->s_inode_list_lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#9 irq_context: 0 (wq_completion)events free_ipc_work proc_inum_ida.xa_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &mapping->i_private_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_IPGRE#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx irq_context: 0 sb_writers#7 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 tomoyo_ss pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&data->suspend_work) irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->unregister_lock irq_context: 0 hci_dev_list_lock irq_context: 0 (work_completion)(&hdev->power_on) irq_context: 0 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (work_completion)(&hdev->reenable_adv_work) irq_context: 0 &hdev->cmd_sync_work_lock irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->interleave_scan)->work) irq_context: 0 &hdev->req_lock hci_dev_list_lock irq_context: 0 &hdev->req_lock (work_completion)(&hdev->tx_work) irq_context: 0 &hdev->req_lock (work_completion)(&hdev->rx_work) irq_context: 0 &hdev->req_lock &wq->mutex irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &wq->mutex &pool->lock irq_context: 0 &hdev->req_lock &wq->mutex &x->wait#10 irq_context: 0 &hdev->req_lock &hdev->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu pool_lock#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &base->lock irq_context: 0 &xt[i].mutex &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &n->list_lock irq_context: 0 nfc_devlist_mutex &n->list_lock &c->lock irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu stock_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock stock_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &base->lock irq_context: 0 &hdev->req_lock &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex &sem->wait_lock irq_context: 0 &data->open_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 key#21 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (console_sem).lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx (wq_completion)nfc15_nci_tx_wq#3 &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx (wq_completion)nfc15_nci_rx_wq#3 irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 misc_mtx (wq_completion)nfc15_nci_rx_wq#3 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx (wq_completion)nfc15_nci_rx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#334 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#334 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 misc_mtx (wq_completion)nfc15_nci_cmd_wq#3 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#4 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#4 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 &hdev->req_lock &hdev->lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#2 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#2 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#333 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#4 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#333 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#3 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#5 &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#5 irq_context: 0 &sb->s_type->i_mutex_key#10 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx quarantine_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#33 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#33 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#33 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#31 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc16_nci_rx_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#335 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#20 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#335 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#335 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1666 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1666 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1666 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1666 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1665 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1665 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1665 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#335 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#334 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#334 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#334 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1665 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1665 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#334 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1665 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#71 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#71 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#71 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#71 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#71 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#493 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#493 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#493 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#493 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#493 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#493 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#493 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#493 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#101 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#101 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#101 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#101 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#101 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 pool_lock#2 irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 namespace_sem &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#100 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1949 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1949 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1949 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1949 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1945 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1940 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#494 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#494 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#494 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#494 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#494 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#34 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#34 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#34 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#34 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#32 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#21 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1667 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1667 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1667 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1667 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1666 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1666 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1666 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1950 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1950 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1950 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock &____s->seqcount#2 irq_context: 0 &hdev->req_lock &hdev->lock &____s->seqcount irq_context: 0 &hdev->req_lock tk_core.seq.seqcount irq_context: 0 &hdev->req_lock hci_sk_list.lock irq_context: 0 &hdev->req_lock &list->lock#7 irq_context: 0 &hdev->req_lock (work_completion)(&hdev->cmd_work) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &hdev->lock irq_context: 0 &hdev->lock fs_reclaim irq_context: 0 &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &hdev->lock pool_lock#2 irq_context: 0 &hdev->lock tk_core.seq.seqcount irq_context: 0 &hdev->lock hci_sk_list.lock irq_context: 0 &hdev->lock &obj_hash[i].lock irq_context: 0 hci_sk_list.lock irq_context: 0 (work_completion)(&rfkill->uevent_work) irq_context: 0 (work_completion)(&rfkill->sync_work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#40 irq_context: 0 subsys mutex#40 &k->k_lock irq_context: 0 subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 &rfkill->lock irq_context: 0 uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#74 irq_context: 0 subsys mutex#74 &k->k_lock irq_context: 0 subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xa->xa_lock#17 irq_context: 0 hci_index_ida.xa_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 misc_mtx cpu_hotplug_lock rcu_node_0 irq_context: 0 misc_mtx cpu_hotplug_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_NETROM irq_context: 0 misc_mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 rtnl_mutex __ip_vs_mutex irq_context: 0 rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex flowtable_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &ul->lock#2 irq_context: 0 rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex &tbl->lock krc.lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: softirq net/wireless/reg.c:533 irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[1] irq_context: 0 &hdev->lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#469 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nn->netlink_tap_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock quarantine_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#53 irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx quarantine_lock irq_context: 0 put_task_map-wait-type-override#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1499 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1720 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_query_lock irq_context: 0 rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 rtnl_mutex &tbl->lock &c->lock irq_context: 0 rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 rtnl_mutex bpf_devs_lock irq_context: 0 rtnl_mutex &hwstats->hwsdev_list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 rtnl_mutex &ul->lock irq_context: 0 rtnl_mutex &net->xdp.lock irq_context: 0 rtnl_mutex mirred_list_lock irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_report_lock irq_context: 0 rtnl_mutex rcu_read_lock sysctl_lock irq_context: 0 rtnl_mutex rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock irq_context: 0 rtnl_mutex rename_lock.seqcount irq_context: 0 rtnl_mutex rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 rtnl_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock pool_lock#2 irq_context: 0 rtnl_mutex &s->s_inode_list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#9 irq_context: 0 rtnl_mutex &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &pnn->pndevs.lock irq_context: 0 rtnl_mutex &pnn->routes.lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 rtnl_mutex deferred_probe_mutex irq_context: 0 rtnl_mutex device_links_lock irq_context: 0 rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &p->lock &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock irq_context: 0 dev_base_lock irq_context: 0 lweventlist_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#114 irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1934 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_NETROM#2 irq_context: 0 nfc_devlist_mutex quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mp->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_LOOPBACK irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#95 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock init_lock irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 misc_mtx pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pcpu_alloc_mutex.wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock key irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->forw_bcast_list_lock irq_context: 0 sb_writers#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&peer->timer_persistent_keepalive) &c->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1910 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1910 irq_context: 0 &rq->__lock &obj_hash[i].lock irq_context: 0 &rq->__lock &base->lock irq_context: 0 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1908 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock sb_writers#3 &____s->seqcount#2 irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu key irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work pool_lock#2 irq_context: 0 rtnl_mutex stack_depot_init_mutex &rq->__lock irq_context: 0 rtnl_mutex stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx fs_reclaim &rq->__lock irq_context: 0 misc_mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 udc_lock udc_lock.wait_lock irq_context: 0 udc_lock.wait_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu irq_context: 0 &dev->lock console_owner_lock irq_context: 0 &hdev->lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1525 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock pool_lock irq_context: softirq (&dum_hcd->timer) &queue->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1963 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 netdev_unregistering_wq.lock irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex quirk_mutex irq_context: 0 sb_internal remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_node_0 irq_context: 0 &dev->mutex remove_cache_srcu irq_context: 0 &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex remove_cache_srcu &c->lock irq_context: 0 &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1958 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1680 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1680 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1679 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1679 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1968 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1968 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: hardirq|softirq &fq->mq_flush_lock quarantine_lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1964 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1959 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1681 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1681 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1680 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1680 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1680 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1680 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#508 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 quarantine_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem key irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem percpu_counters_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_lock stock_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#508 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#506 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#506 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1969 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1969 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1969 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1969 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1965 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1960 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1970 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1970 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1966 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 sb_writers#3 &____s->seqcount#2 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem (console_sem).lock irq_context: 0 cgroup_threadgroup_rwsem kernfs_pr_cont_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &n->list_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci5#3 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 cgroup_threadgroup_rwsem kernfs_pr_cont_lock kernfs_rename_lock irq_context: 0 cgroup_threadgroup_rwsem kernfs_pr_cont_lock (console_sem).lock irq_context: 0 cgroup_threadgroup_rwsem console_lock console_srcu console_owner_lock irq_context: 0 cgroup_threadgroup_rwsem console_lock console_srcu console_owner irq_context: 0 cgroup_threadgroup_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 cgroup_threadgroup_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &cfs_rq->removed.lock irq_context: 0 kn->active#4 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1934 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1934 irq_context: softirq (&dum_hcd->timer) &x->wait#27 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->can.stattimer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem vmap_area_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem recent_lock irq_context: 0 rtnl_mutex stock_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: softirq (&dum_hcd->timer) &queue->lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock &rq->__lock irq_context: 0 &ndev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock irq_context: softirq (&ndev->rs_timer) batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &n->list_lock irq_context: 0 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &disk->open_mutex &lock->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 quarantine_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_node_0 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1525 &rq->__lock irq_context: 0 &dev->mutex quarantine_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1961 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 &f->f_pos_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1961 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#351 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1961 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1682 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1682 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 misc_mtx rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &rcu_state.expedited_wq irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#14 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &c->lock irq_context: 0 &dev_instance->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev_instance->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1450 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1932 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex.wait_lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &cfs_rq->removed.lock irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock kfence_freelist_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1682 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#344 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#344 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1682 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock subsys mutex#76 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->k_lock klist_remove_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 cgroup_threadgroup_rwsem rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pidmap_lock pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock &base->lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_file_kn_lock &base->lock &obj_hash[i].lock irq_context: softirq (&cfile->notify_timer) irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&cfile->notify_timer) cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mirred_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_report_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex target_list_lock irq_context: 0 fs_reclaim pcpu_lock irq_context: 0 fs_reclaim percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_NONE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1472 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1472 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1473 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1473 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1473 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu irq_context: 0 nfc_devlist_mutex remove_cache_srcu quarantine_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &c->lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &rq->__lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex remove_cache_srcu &n->list_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem napi_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netdev_unregistering_wq.lock irq_context: 0 lock pidmap_lock &____s->seqcount#2 irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex rcu_node_0 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#343 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1681 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#343 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1681 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#3 quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1716 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1717 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1907 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#114 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu pool_lock#2 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem key irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_SIT#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_SIT irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1473 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1330 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 key#6 irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &n->list_lock irq_context: softirq (&ndev->rs_timer) &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1927 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1480 irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1336 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex triggers_list_lock &rq->__lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &rq->__lock irq_context: 0 &xt[i].mutex &base->lock irq_context: 0 &xt[i].mutex &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1493 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1493 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1494 irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex batched_entropy_u8.lock irq_context: 0 &tsk->futex_exit_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex kfence_freelist_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem stock_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem key irq_context: 0 &dev->mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1510 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1510 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1516 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1516 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: softirq (&timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1651 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 &base->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1651 &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1651 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex &rq->__lock irq_context: 0 loop_validate_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1650 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1618 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1618 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#115 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#115 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#115 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#115 &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock stock_lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1650 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1650 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1650 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 udc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#482 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#482 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#482 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#482 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem stock_lock irq_context: 0 rtnl_mutex dev_addr_sem key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1935 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1935 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss stock_lock irq_context: 0 tomoyo_ss pcpu_lock stock_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1950 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1946 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1946 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1946 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1941 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#336 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#336 &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1935 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1935 irq_context: 0 sb_writers#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 nfc_devlist_mutex kn->active#4 irq_context: 0 nfc_devlist_mutex kn->active#4 &root->deactivate_waitq irq_context: 0 nfc_devlist_mutex kn->active#4 &rq->__lock irq_context: 0 nfc_devlist_mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#116 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#116 irq_context: 0 &root->deactivate_waitq irq_context: 0 &root->deactivate_waitq &p->pi_lock irq_context: 0 &root->deactivate_waitq &p->pi_lock &rq->__lock irq_context: 0 &root->deactivate_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex key irq_context: 0 misc_mtx nfc_devlist_mutex pcpu_lock irq_context: 0 misc_mtx nfc_devlist_mutex percpu_counters_lock irq_context: 0 misc_mtx nfc_devlist_mutex pcpu_lock stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx pcpu_alloc_mutex.wait_lock irq_context: 0 &dir->lock#2 irq_context: 0 &dir->lock#2 &obj_hash[i].lock irq_context: 0 &dir->lock#2 pool_lock#2 irq_context: 0 &dir->lock &obj_hash[i].lock irq_context: 0 &dir->lock pool_lock#2 irq_context: 0 css_set_lock cgroup_file_kn_lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock tasklist_lock irq_context: 0 (wq_completion)netns irq_context: 0 (wq_completion)netns net_cleanup_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->nsid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &wg->socket_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fsnotify_mark_srcu &____s->seqcount#2 irq_context: 0 &fsnotify_mark_srcu &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock smack_known_lock.wait_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu &c->lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu pool_lock#2 irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &obj_hash[i].lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1933 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work quarantine_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback proc_inum_ida.xa_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &table->hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &fsnotify_mark_srcu irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1651 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1936 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &group->mark_mutex &lock->wait_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1936 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1937 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock key irq_context: 0 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1549 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1400 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1400 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &tn->node_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ebt_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 rtnl_mutex dev_addr_sem quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#509 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &meta->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss stock_lock irq_context: 0 &dev->mutex uevent_sock_mutex stock_lock irq_context: softirq rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1720 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex quarantine_lock irq_context: softirq (&peer->timer_retransmit_handshake) irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &meta->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 sched_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#509 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#107 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#107 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &meta->lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &n->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &n->list_lock &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &base->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex irq_context: 0 sb_writers &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &f->f_pos_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 sb_writers#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1720 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1719 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock key irq_context: 0 sb_writers#3 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#3 sb_internal kfence_freelist_lock irq_context: 0 sb_writers#3 sb_internal &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1937 &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex ebt_mutex.wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#11 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex klist_remove_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netns_bpf_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock deferred_probe_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock sysfs_symlink_target_lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#4 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) pool_lock#2 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) quarantine_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) remove_cache_srcu irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci5#4 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#4 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#37 irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#38 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#507 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#107 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#74 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#107 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#107 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#106 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#507 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1971 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1971 &rq->__lock irq_context: 0 &xt[i].mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &x->wait#2 irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_probe_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->cells_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->cells_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)afs irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &net->cells_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &(&net->fs_lock)->lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->incoming_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &call->waitq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &call->notify_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (rxrpc_call_limiter).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->recvmsg_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (&call->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &list->lock#17 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)kafsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &local->services_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krxrpcd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rlock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rxnet->local_mutex irq_context: 0 (&local->client_conn_reap_timer) irq_context: 0 &rxnet->conn_lock irq_context: 0 &table->hash[i].lock irq_context: 0 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-clock-AF_INET6 irq_context: 0 &list->lock#18 irq_context: 0 &xt[i].mutex quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1971 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1646 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex (work_completion)(&data->gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ovs_net->dp_notify_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &srv->idr_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1646 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock irq_context: 0 sb_writers#5 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1971 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1967 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1962 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1683 irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1287 irq_context: 0 rtnl_mutex lweventlist_lock batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu stock_lock irq_context: 0 remove_cache_srcu key irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1722 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 rcu_node_0 irq_context: softirq (&dum_hcd->timer) &queue->lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1804 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1559 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#387 &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#387 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &base->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#389 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#389 &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &rq->__lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#36 input_pool.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc7_nci_rx_wq#47 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: softirq (&timer) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1514 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 tomoyo_ss &base->lock irq_context: 0 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#405 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1523 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1523 &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &dev_instance->mutex remove_cache_srcu &rq->__lock irq_context: 0 &dev_instance->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &obj_hash[i].lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fn->fou_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem key irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem pcpu_lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->sync_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (inetaddr_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem inet6addr_chain.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wdev->pmsr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &rq->__lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1532 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1532 &rq->__lock irq_context: 0 &vma->vm_lock->lock key irq_context: 0 &vma->vm_lock->lock pcpu_lock irq_context: 0 &vma->vm_lock->lock percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &____s->seqcount irq_context: 0 &vma->vm_lock->lock pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1532 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1752 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1752 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1752 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1752 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&ifibss->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&wdev->disconnect_wk) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&wdev->pmsr_free_wk) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ret->b_state_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1753 &rq->__lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1753 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#413 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#413 &rq->__lock irq_context: 0 &child->perf_event_mutex &rq->__lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#413 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#413 irq_context: 0 misc_mtx wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 misc_mtx wq_pool_mutex &rq->__lock irq_context: 0 misc_mtx wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#413 irq_context: 0 &p->lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &p->lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu quarantine_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &c->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &obj_hash[i].lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&local->dynamic_ps_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->i_private_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->dev_wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_base_lock &xa->xa_lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx bpf_devs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mirred_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_report_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnn->pndevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnn->routes.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx target_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->k_lock klist_remove_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#413 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#287 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#414 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &n->list_lock irq_context: 0 &vma->vm_lock->lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&local->restart_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rdev->wiphy.mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rfkill->uevent_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rfkill->sync_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rfkill->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem triggers_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem leds_list_lock irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock fs_reclaim &rq->__lock irq_context: 0 cb_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &meta->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex nf_sockopt_mutex.wait_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &group->mark_mutex &n->list_lock irq_context: 0 &group->mark_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1542 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->conn_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->event_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&rdev->dfs_update_channels_wk)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&rdev->background_cac_done_wk)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->destroy_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->propagate_radar_detect_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->propagate_cac_done_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->mgmt_registrations_update_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->background_cac_abort_wk) irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&local->sta_cleanup) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 bt_proto_lock &n->list_lock irq_context: 0 bt_proto_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->dat.work)->timer irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->bla.work)->timer irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &hash->list_locks[i] irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) key#19 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &entry->crc_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 &dir->lock &meta->lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dir->lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1779 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1915 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1915 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &root->kernfs_rwsem irq_context: softirq &(&net->ipv6.addr_chk_work)->timer irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex lweventlist_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1288 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &type->s_umount_key#47 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock kfence_freelist_lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex key irq_context: 0 &xt[i].mutex pcpu_lock irq_context: 0 &xt[i].mutex percpu_counters_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1683 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1682 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1682 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1682 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1682 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#510 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#510 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#508 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#508 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#345 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#345 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#344 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#344 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#344 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#344 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#108 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#108 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1972 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1972 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1972 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1972 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1968 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1968 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1968 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1963 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1963 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1963 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1684 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1684 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1683 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#346 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1683 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1683 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#346 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem rdma_nets.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem devices_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nlk->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &base->lock irq_context: 0 &p->lock &cfs_rq->removed.lock irq_context: 0 &p->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &pcp->lock &zone->lock irq_context: 0 &fsnotify_mark_srcu key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1903 irq_context: 0 &fsnotify_mark_srcu pcpu_lock irq_context: 0 &fsnotify_mark_srcu percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &fsnotify_mark_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1629 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1629 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &xa->xa_lock#18 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1629 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1629 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &app->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->join_timer)#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->periodic_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &list->lock#11 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->join_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &app->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &list->lock#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&priv->scan_result)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&priv->connect)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&hsr->prune_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&hsr->announce_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1628 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1628 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1628 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1628 irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) &rq->__lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy66 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) quarantine_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1911 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1911 irq_context: 0 &data->open_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex j1939_netdev_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &vlan_netdev_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &batadv_netdev_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_LOOPBACK#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 &sch->q.lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#54 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#54 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1889 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1889 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1619 irq_context: 0 &p->lock &of->mutex kn->active#4 kfence_freelist_lock irq_context: 0 sb_writers#3 &rq->__lock cpu_asid_lock irq_context: 0 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 vlan_ioctl_mutex rtnl_mutex.wait_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 vlan_ioctl_mutex &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex batched_entropy_u8.lock irq_context: 0 misc_mtx nfc_devlist_mutex kfence_freelist_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1909 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1889 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1889 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fsnotify_mark_srcu irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &obj_hash[i].lock pool_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &list->lock#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &r->consumer_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->socket_update_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->arp_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->slave_arr_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->hello_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->topology_change_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->tcn_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_other_query.delay_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_other_query.delay_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 krc.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 sb_internal &c->lock irq_context: 0 sb_writers &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2014 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#55 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 file_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1559 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_PACKET &rnp->exp_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount#2 irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#327 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#327 &rq->__lock irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex fs_reclaim &rq->__lock irq_context: 0 &data->open_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq rcu_callback &n->list_lock irq_context: softirq rcu_callback &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2009 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1730 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 misc_mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &rnp->exp_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 &dev->mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[0] irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#15 irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock (wq_completion)hci3#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy62 irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy62 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy61 irq_context: 0 (wq_completion)hci3#9 irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex raw_notifier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bcm_notifier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex isotp_notifier_lock irq_context: 0 &root->kernfs_rwsem key irq_context: 0 &root->kernfs_rwsem pcpu_lock irq_context: 0 &root->kernfs_rwsem percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 &root->kernfs_rwsem rcu_node_0 irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 krc.lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock &c->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#69 irq_context: 0 (wq_completion)wg-kex-wg1#69 &rq->__lock irq_context: 0 &fc->uapi_mutex &c->lock irq_context: 0 &fc->uapi_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex napi_hash_lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex req_lock irq_context: 0 &type->i_mutex_dir_key/1 irq_context: 0 &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#11 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tap_major->minor_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&port->bc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&port->wq) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex rcu_node_0 irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events &rq->__lock irq_context: 0 (wq_completion)bat_events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1910 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1636 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &k->k_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1636 &rq->__lock irq_context: 0 rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#60 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#59 irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx pool_lock#2 irq_context: 0 remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 krc.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers#3 &fsnotify_mark_srcu rcu_node_0 irq_context: 0 misc_mtx rfkill_global_mutex &____s->seqcount#2 irq_context: 0 cb_lock remove_cache_srcu irq_context: 0 cb_lock remove_cache_srcu quarantine_lock irq_context: 0 cb_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1620 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &meta->lock irq_context: 0 drm_unplug_srcu irq_context: 0 &dev->event_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1937 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 &dev_instance->mutex vicodec_core:1851:(hdl)->_lock &rq->__lock irq_context: 0 &dev_instance->mutex vicodec_core:1851:(hdl)->_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1937 irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#346 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1683 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1683 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1683 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#346 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#346 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#346 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#345 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#345 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#109 irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &lruvec->lru_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem &ei->i_raw_lock irq_context: 0 misc_mtx stock_lock irq_context: 0 misc_mtx key irq_context: 0 misc_mtx pcpu_lock irq_context: 0 misc_mtx percpu_counters_lock irq_context: 0 misc_mtx pcpu_lock stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &n->list_lock irq_context: 0 (wq_completion)hci1#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&slave->notify_work)->work) irq_context: 0 tasklist_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 sb_internal jbd2_handle key#4 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->ipsec_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex crngs.lock irq_context: 0 sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &cfs_rq->removed.lock irq_context: 0 &xa->xa_lock#5 &c->lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 nf_hook_mutex &rq->__lock irq_context: 0 nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock irq_context: 0 &fsnotify_mark_srcu &rq->__lock cpu_asid_lock irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#529 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1935 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1930 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1653 irq_context: 0 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#109 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#109 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#108 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1685 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1685 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1685 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1685 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1684 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1684 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#347 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#347 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#347 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#346 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#346 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1686 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1686 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1686 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1686 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1685 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1685 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1685 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1685 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1685 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1685 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#348 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#348 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#347 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#347 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1687 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1687 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1687 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1687 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#110 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1686 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1686 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1686 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#349 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#349 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#349 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#110 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#110 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#109 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#109 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#109 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#349 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#348 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#348 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#348 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#77 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#77 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#77 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#77 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#77 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1687 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1687 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1687 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1687 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1687 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1687 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#350 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#350 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#349 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#349 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#349 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#349 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#511 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#511 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#509 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#509 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1973 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1973 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1969 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1964 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1974 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1974 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1974 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1974 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1970 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1965 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1689 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1689 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1688 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1688 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1690 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1690 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1690 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1690 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1689 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1689 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1689 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1689 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1689 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1689 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#111 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#111 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#111 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#111 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#111 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1691 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1691 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1691 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1691 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1690 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1690 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1690 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1690 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#112 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#112 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#112 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#112 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#112 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#112 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#111 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#512 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#512 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#510 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#510 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#351 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#351 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#351 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#351 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#350 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#350 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1975 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1975 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1971 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1971 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1971 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1966 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1976 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1976 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1972 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1967 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1692 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1692 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1692 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1692 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wg->socket_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock key#22 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1691 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1691 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1691 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1691 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#511 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#511 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#511 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1977 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1977 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1977 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1973 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1968 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1978 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1978 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1978 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1978 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1974 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1969 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1693 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1693 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1692 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1692 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1979 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1979 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1979 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1979 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1979 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1975 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1970 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1694 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1694 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1694 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1693 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1693 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1980 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1980 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1980 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1980 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1976 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1971 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1971 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1695 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1695 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1695 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1695 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1694 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1694 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1694 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &table->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock napi_hash_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#514 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#514 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#514 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#512 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#512 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &x->wait#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1804 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1804 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#40 irq_context: 0 misc_mtx rfkill_global_mutex &n->list_lock irq_context: 0 misc_mtx rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &fsnotify_mark_srcu fs_reclaim &rq->__lock irq_context: 0 &fsnotify_mark_srcu fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#512 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#512 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1981 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1981 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1981 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1653 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock cpu_asid_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &list->lock#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback key#22 irq_context: 0 (wq_completion)events reg_work rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_mayday_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &x->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &r->consumer_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events reg_work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1800 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1981 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1977 irq_context: 0 tomoyo_ss remove_cache_srcu &meta->lock irq_context: 0 tomoyo_ss remove_cache_srcu kfence_freelist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1972 irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1982 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &r->producer_lock#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)rcu_gp &rq->__lock irq_context: 0 (wq_completion)rcu_gp &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 nf_hook_mutex &n->list_lock irq_context: 0 nf_hook_mutex &n->list_lock &c->lock irq_context: 0 nf_nat_proto_mutex &n->list_lock irq_context: 0 nf_nat_proto_mutex &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 uevent_sock_mutex &n->list_lock irq_context: 0 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &base->lock irq_context: 0 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 quarantine_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vmap_purge_lock &rq->__lock irq_context: 0 vmap_purge_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1982 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1982 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1982 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1978 irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1978 &rq->__lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1978 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1973 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1973 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1973 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1696 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1696 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock cpu_asid_lock irq_context: 0 &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1695 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1695 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1695 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1695 irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#515 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#515 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#513 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#513 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#352 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#352 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->gw.list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.handler_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#17 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hash->list_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#19 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#19 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#19 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#15 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#15 &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#15 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#15 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#15 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.req_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.roam_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#18 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#352 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#351 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#351 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#58 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_mayday_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hn->hn_lock irq_context: 0 (wq_completion)inet_frag_wq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &obj_hash[i].lock pool_lock irq_context: 0 udc_lock &cfs_rq->removed.lock irq_context: 0 udc_lock &obj_hash[i].lock irq_context: 0 udc_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &this->info_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnetids_ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 &list->lock#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_node_0 irq_context: 0 lock kernfs_idr_lock &n->list_lock irq_context: 0 lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 prog_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#351 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#351 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#351 &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) irq_context: softirq (&peer->timer_send_keepalive) &c->lock irq_context: softirq (&peer->timer_send_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_send_keepalive) &list->lock#14 irq_context: softirq (&peer->timer_send_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_internal &rq->__lock irq_context: 0 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#351 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#113 irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#113 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#113 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle irq_context: softirq (&lapb->t1timer) &lapb->lock &____s->seqcount#2 irq_context: softirq (&lapb->t1timer) &lapb->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex pool_lock#2 irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#78 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#78 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#113 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#78 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1983 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1696 irq_context: 0 &fsnotify_mark_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &dev->lock irq_context: 0 &dev->mutex &queue->lock irq_context: 0 &dev->mutex &queue->lock pool_lock#2 irq_context: 0 &dev->mutex &queue->lock semaphore->lock#2 irq_context: 0 &dev->mutex &udc->connect_lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock &dev->power.lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock &dum_hcd->dum->lock hcd_root_hub_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock irq_context: 0 &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 semaphore->lock#2 irq_context: 0 &queue->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock device_state_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex usbfs_mutex irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#47 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#47 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex kfence_freelist_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &n->list_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &base->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &meta->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock irq_context: softirq (&dum_hcd->timer) &queue->lock pool_lock#2 irq_context: 0 &root->kernfs_rwsem stock_lock irq_context: 0 &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu quarantine_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &c->lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq lock#6 &kcov->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &ndev->req_lock &wq->mutex &rq->__lock irq_context: 0 &ndev->req_lock &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &x->wait#19 irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &base->lock irq_context: 0 &kcov->lock kcov_remote_lock &n->list_lock irq_context: 0 &kcov->lock kcov_remote_lock &n->list_lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1529 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1529 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1530 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1530 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1530 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1531 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1531 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#367 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#367 &rq->__lock irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1532 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1532 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1533 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1526 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1526 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1526 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1526 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1527 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1527 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1527 irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1528 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1533 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1534 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1534 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1534 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1535 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1401 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1614 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1615 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1550 &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1402 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss key irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1615 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1615 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1615 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pcpu_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1462 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1462 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1462 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1462 irq_context: softirq &c->lock batched_entropy_u8.lock irq_context: softirq &c->lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1463 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1463 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 key irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy44 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy43 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1553 irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1553 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1553 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1554 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1405 irq_context: 0 &ep->mtx fs_reclaim &rq->__lock irq_context: 0 &hdev->req_lock (wq_completion)hci4#6 irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 &p->pi_lock irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#21 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#353 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#42 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#41 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#21 irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#41 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1627 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1627 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1627 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1627 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#21 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#21 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex pool_lock#2 irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#24 irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#24 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1676 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1676 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1677 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1677 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1677 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1482 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1482 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1677 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1482 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1483 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1483 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1483 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev_instance->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &n->list_lock irq_context: 0 rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 sk_lock-AF_INET rcu_node_0 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#78 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#113 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex usb_port_peer_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex usb_port_peer_mutex device_state_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_links_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex (&timer.timer) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex (console_sem).lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &hdev->req_lock (wq_completion)hci2#14 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#112 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock pool_lock#2 irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock semaphore->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex (&timer.timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 (console_sem).lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 lock#4 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) init_task.mems_allowed_seq.seqcount irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &base->lock irq_context: 0 &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_iattr_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#37 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#73 irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#116 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#117 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#117 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#117 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET resource_mutex &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET resource_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex remove_cache_srcu irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1983 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1983 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#336 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#335 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#335 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#495 irq_context: 0 &xt[i].mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#495 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#102 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#102 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#102 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#495 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#495 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#495 &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#101 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#495 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#495 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2024 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#36 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 nl_table_wait.lock irq_context: 0 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2024 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2024 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2024 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1552 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1552 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1652 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1652 irq_context: 0 &disk->open_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1652 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1652 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#484 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#484 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1939 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1939 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1939 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1940 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1940 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1938 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &dev->power.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1941 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1928 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1941 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1368 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1368 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1369 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&dum_hcd->timer) &x->wait#27 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &rq->__lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock pool_lock#2 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#36 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &in_dev->mc_tomb_lock irq_context: 0 nfc_devlist_mutex device_links_lock &rq->__lock irq_context: 0 nfc_devlist_mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem xfrm_state_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_state_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &lru->node[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ip6_fl_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: softirq (&peer->timer_persistent_keepalive) init_task.mems_allowed_seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->rules_mod_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->ipv6.ip6_fib_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mrt->ipmr_expire_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&tbl->gc_work)->timer irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_mutex irq_context: 0 (wq_completion)hci1#12 &rq->__lock irq_context: 0 (wq_completion)hci1#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&ipvs->dest_trash_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hashlimit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hashlimit_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem trans_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1939 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1934 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1656 irq_context: 0 &type->i_mutex_dir_key#6 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#6 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#6 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#6 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#6 &xa->xa_lock#5 irq_context: 0 &type->i_mutex_dir_key#6 &xa->xa_lock#5 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &____s->seqcount#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex rcu_read_lock &nf_nat_locks[i] irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1942 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem quarantine_lock irq_context: 0 &dev->mutex udc_lock udc_lock.wait_lock irq_context: 0 &dev->mutex udc_lock &rq->__lock irq_context: 0 &dev->mutex udc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex udc_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#118 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#118 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#118 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#119 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#119 irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount irq_context: 0 uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1983 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1979 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1974 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#516 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#516 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#516 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#516 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET pool_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&dum_hcd->timer) &queue->lock &n->list_lock irq_context: softirq (&dum_hcd->timer) &queue->lock &n->list_lock &c->lock irq_context: softirq (&peer->timer_persistent_keepalive) &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem tcp_metrics_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.policy_hash_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_policy_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.state_hash_work) irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu pool_lock#2 irq_context: softirq (&ndev->rs_timer) kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#4 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem genl_sk_destructing_waitq.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) crngs.lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock udc_lock.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock.wait_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &p->pi_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &p->pi_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 krc.lock &base->lock irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1371 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1371 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1371 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1371 irq_context: 0 &uhid->devlock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock batched_entropy_u8.lock irq_context: 0 &p->lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#16 irq_context: 0 &p->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mount_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1559 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1558 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1558 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#514 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1474 irq_context: 0 videodev_lock &rq->__lock irq_context: 0 videodev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#359 irq_context: 0 (wq_completion)events free_ipc_work stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock remove_cache_srcu &rq->__lock irq_context: 0 cb_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#36 lock kernfs_idr_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1475 irq_context: 0 rtnl_mutex team->team_lock_key#36 &root->kernfs_rwsem irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1475 irq_context: 0 cb_lock remove_cache_srcu pool_lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1943 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1659 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1476 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1476 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1332 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1332 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: softirq (&peer->timer_persistent_keepalive) batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &____s->seqcount irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1476 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1659 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1658 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#487 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#332 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1337 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &c->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1337 irq_context: softirq (&pool->idle_timer) irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#488 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1486 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1487 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1487 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1343 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1487 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1487 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1489 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1489 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1489 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &rq->__lock irq_context: 0 &dev->mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_attach_mutex &rq->__lock irq_context: 0 wq_pool_attach_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex quarantine_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 &dev->mutex pinctrl_list_mutex &rq->__lock irq_context: 0 &dev->mutex pinctrl_list_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1494 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1350 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1350 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1494 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1494 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->i_private_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1495 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1495 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1495 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1496 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1496 &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1496 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1497 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1353 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#361 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#361 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#361 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1354 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1499 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1499 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1500 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1500 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1355 irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1355 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1355 irq_context: 0 &uhid->devlock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1355 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1500 irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1500 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1500 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1501 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1501 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#119 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#120 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1356 irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock &n->list_lock irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1357 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1504 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#40 &rq->__lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 irq_context: 0 &dev->mutex subsys mutex#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work ima_keys_lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tty_mutex &n->list_lock irq_context: 0 tty_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#488 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#488 irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &lock->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1509 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1509 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex quarantine_lock irq_context: 0 sb_writers#5 rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1510 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_node_0 irq_context: 0 rtnl_mutex &xa->xa_lock#4 &c->lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &n->list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1510 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1372 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1372 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#346 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#346 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#247 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1374 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1374 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1238 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1239 irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1382 irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#488 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1365 irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1662 irq_context: 0 rtnl_mutex dev_addr_sem pcpu_lock irq_context: 0 rtnl_mutex dev_addr_sem percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1662 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#36 lweventlist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#333 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#333 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) rcu_node_0 irq_context: 0 tty_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1511 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1511 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1511 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu irq_context: 0 &xt[i].mutex remove_cache_srcu pool_lock#2 irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex kn->active#4 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex kn->active#4 &root->deactivate_waitq irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &n->list_lock &c->lock irq_context: 0 &dev->mutex kn->active#4 &rq->__lock irq_context: 0 &dev->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &n->list_lock &c->lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 lock map_idr_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &uhid->qlock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1386 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1386 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1386 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &uhid->qlock &uhid->waitq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1386 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1387 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1387 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1387 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1387 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1387 &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex team->team_lock_key#36 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#36 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 (console_sem).lock irq_context: 0 lock map_idr_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#36 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#36 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#36 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 fs_reclaim irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1387 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1246 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1246 &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 &uhid->devlock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1387 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1387 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#349 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#250 irq_context: 0 dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1388 irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1388 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1389 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1389 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1389 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_node_0 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#514 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#514 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1511 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1367 irq_context: 0 tty_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1511 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1511 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1511 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1512 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1512 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) &x->wait#19 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &____s->seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 udc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbinfo->stat_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &queue->lock &c->lock irq_context: 0 &dev->mutex &queue->lock &n->list_lock irq_context: 0 &dev->mutex &queue->lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&timer) rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1521 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1521 irq_context: 0 tasklist_lock &base->lock irq_context: 0 tasklist_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1522 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1375 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1522 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1522 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1522 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1522 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1523 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1523 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock &rq->__lock irq_context: 0 lock map_idr_lock pool_lock#2 irq_context: 0 map_idr_lock irq_context: 0 map_idr_lock &obj_hash[i].lock irq_context: 0 map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#53 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1523 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1523 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1523 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1524 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1524 irq_context: 0 &ep->mtx stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1524 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1377 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1389 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#350 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#351 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1394 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1397 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &pcp->lock &zone->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1397 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1400 irq_context: 0 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1400 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1401 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1261 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1377 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1403 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1377 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1378 irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 kn->active#14 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#11 &n->list_lock irq_context: 0 kn->active#11 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1378 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1378 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 pool_lock#2 irq_context: 0 file_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1378 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1378 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1378 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1379 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#365 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#365 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#365 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#365 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#365 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#365 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#365 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#365 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &____s->seqcount#2 irq_context: 0 nfc_devlist_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1524 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#260 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#333 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex batched_entropy_u32.lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex pcpu_alloc_mutex irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#12 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#36 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#36 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1662 irq_context: 0 rtnl_mutex team->team_lock_key#36 &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock kfence_freelist_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#33 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#490 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex bpf_devs_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rcu_state.expedited_wq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex batched_entropy_u8.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[0] irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 sb_writers#4 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#4 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#48 irq_context: 0 (wq_completion)wg-kex-wg0#48 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#47 irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1415 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1415 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1415 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start stock_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start key irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#51 pool_lock#2 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#51 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#51 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#384 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#384 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy32 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1288 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1288 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#384 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#384 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#514 &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#353 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#353 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#353 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#353 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#384 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#352 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#384 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#384 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#40 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#40 irq_context: 0 &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#40 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#40 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#267 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#267 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#267 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#267 irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock pool_lock#2 irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)nfc5_nci_rx_wq#352 &rq->__lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &n->list_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &rq->__lock irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1566 &rq->__lock irq_context: 0 &pipe->mutex/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#254 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1566 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1566 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1567 irq_context: 0 sk_lock-AF_ROSE irq_context: 0 sk_lock-AF_ROSE slock-AF_ROSE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1429 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 slock-AF_ROSE irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1430 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1430 &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1430 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1431 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1431 irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#30 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#30 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock pool_lock#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1288 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1288 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#29 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1431 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1431 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1431 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1432 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1432 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#15 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1450 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1453 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1453 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1456 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1457 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1567 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1567 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1567 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1567 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_node_0 irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1466 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1468 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1324 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1468 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1468 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1469 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1469 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1469 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1469 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1568 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1568 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1569 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#492 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#492 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#492 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#492 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#70 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1664 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1928 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1928 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1664 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1663 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1665 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1665 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1664 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1664 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1664 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#47 rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#47 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#36 irq_context: 0 &type->s_umount_key#47 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 &type->s_umount_key#47 rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#18 irq_context: 0 (wq_completion)wg-kex-wg0#36 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#18 irq_context: 0 (wq_completion)wg-kex-wg0#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#9 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#9 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#35 irq_context: 0 tty_mutex &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#529 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#352 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#352 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#515 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#515 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#517 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#354 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#114 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#114 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#114 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#113 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1697 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1697 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#518 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#36 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1569 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1419 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#518 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#516 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1419 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1569 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1569 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1420 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1420 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1570 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1570 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &rq->__lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &queue->lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1571 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1571 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1421 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1421 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1571 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1571 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1571 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1572 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1572 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 rtnl_mutex bus_type_sem &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bus_type_sem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex bus_type_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex bus_type_sem pool_lock#2 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 rtnl_mutex &rnp->exp_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[2] irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1572 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1572 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1573 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#370 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#370 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#370 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1575 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1575 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1576 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock rcu_read_lock &n->list_lock irq_context: 0 cb_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &n->list_lock irq_context: 0 sb_writers#8 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1576 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1425 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1425 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1576 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1576 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1576 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1576 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1577 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1577 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1577 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1577 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET percpu_counters_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &rnp->exp_wq[0] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1514 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1514 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1514 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1514 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#57 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1727 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#390 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1288 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#516 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#516 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#3 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#3 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 sb_internal &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1289 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#58 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#58 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &meta->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#515 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#515 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock udc_lock udc_lock.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1728 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1506 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1432 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1552 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1552 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_node_0 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rcu_state.expedited_wq irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1577 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1577 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &meta->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1426 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1577 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1577 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1577 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1577 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1578 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &meta->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1578 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1427 irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1578 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1578 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1579 irq_context: 0 &dev->mutex uevent_sock_mutex key irq_context: 0 &dev->mutex uevent_sock_mutex pcpu_lock irq_context: 0 &dev->mutex uevent_sock_mutex percpu_counters_lock irq_context: 0 &dev->mutex uevent_sock_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1579 &rq->__lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 kn->active#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1579 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1579 &rq->__lock irq_context: softirq (&lapb->t1timer) &lapb->lock &list->lock#20 irq_context: softirq &list->lock#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1579 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1579 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex batched_entropy_u8.lock irq_context: 0 misc_mtx &dev->mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1580 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1429 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1429 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1429 irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1580 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1580 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1580 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1581 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1581 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1430 &rq->__lock irq_context: 0 sk_lock-AF_VSOCK &rq->__lock irq_context: 0 sk_lock-AF_VSOCK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1431 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1431 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &base->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &dum_hcd->dum->lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex (&timer.timer) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pcpu_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1432 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1582 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1582 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1583 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1583 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1583 irq_context: 0 &dev->mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 &dev->mutex uevent_sock_mutex &meta->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1584 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1730 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#355 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1730 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#60 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1584 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_node_0 irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1584 irq_context: 0 &f->f_pos_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1584 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1585 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1585 irq_context: 0 &f->f_pos_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#392 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#392 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1729 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#59 irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1620 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1620 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1621 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1621 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1468 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1628 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1628 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1622 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1622 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1622 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1469 irq_context: 0 ebt_mutex &cfs_rq->removed.lock irq_context: 0 ebt_mutex &obj_hash[i].lock irq_context: 0 ebt_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1624 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1624 &rq->__lock irq_context: 0 sb_writers#3 sb_internal quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1624 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &____s->seqcount irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#120 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1633 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1633 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1633 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1635 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1640 irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1435 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1480 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1480 &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &base->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#120 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev_instance->mutex remove_cache_srcu &c->lock irq_context: 0 &dev_instance->mutex remove_cache_srcu &n->list_lock irq_context: 0 &dev_instance->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &p->pi_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) quarantine_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1651 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#330 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#483 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#483 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1636 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#355 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#60 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#355 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1663 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1663 irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#24 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#47 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1674 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1674 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &meta->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1934 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#22 irq_context: 0 sb_writers#5 stock_lock irq_context: 0 sb_writers#5 key irq_context: 0 lock pidmap_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem quarantine_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1684 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pcpu_lock stock_lock irq_context: 0 &uhid->devlock semaphore->lock#3 irq_context: 0 sb_writers#3 &mm->mmap_lock &info->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_node_0 irq_context: 0 &xt[i].mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1696 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1696 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1697 irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &bgl->locks[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1697 &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock stock_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss quarantine_lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &base->lock irq_context: 0 &type->i_mutex_dir_key#4 &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1703 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1703 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#32 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1711 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1711 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1711 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)nfc8_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1721 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1721 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1721 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1721 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1721 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1720 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1720 &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1720 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#385 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#385 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#120 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#385 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#385 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#385 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#385 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#385 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#385 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#385 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#385 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1500 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1500 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1500 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1500 irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#47 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 &type->s_umount_key#47 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1500 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1500 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1722 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xa->xa_lock#9 &n->list_lock irq_context: 0 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#11 irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#268 irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#268 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#268 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1722 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1721 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#52 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#52 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#52 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET stock_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#268 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#268 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#52 irq_context: 0 &xt[i].mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 remove_cache_srcu pcpu_lock irq_context: 0 remove_cache_srcu percpu_counters_lock irq_context: 0 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 &xt[i].mutex batched_entropy_u8.lock irq_context: 0 &xt[i].mutex kfence_freelist_lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &meta->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock kfence_freelist_lock irq_context: softirq (&dum_hcd->timer) &queue->lock kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_owner irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 smack_known_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 sb_internal jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 &xa->xa_lock#9 &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#121 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#4 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#270 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#61 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#61 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#61 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#61 irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#61 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: softirq (&timer) &obj_hash[i].lock irq_context: softirq (&timer) &base->lock irq_context: softirq (&timer) &base->lock &obj_hash[i].lock irq_context: softirq (&timer) rcu_read_lock pool_lock#2 irq_context: softirq (&timer) rcu_read_lock &c->lock irq_context: softirq (&timer) rcu_read_lock &n->list_lock irq_context: softirq (&timer) rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&timer) &txlock irq_context: softirq (&timer) &txlock &list->lock#3 irq_context: softirq (&timer) &txwq irq_context: softirq (&timer) &txwq &p->pi_lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh &list->lock#12 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM (console_sem).lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key nr_node_list_lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key pool_lock#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1741 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1741 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1529 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock cpu_asid_lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1536 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1756 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1756 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1756 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1538 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1538 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &hsr->list_lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1779 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1509 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &child->perf_event_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &xa->xa_lock#5 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock cpu_asid_lock irq_context: 0 &hdev->req_lock (wq_completion)hci2#12 irq_context: 0 (wq_completion)hci0#9 irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1779 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 nf_sockopt_mutex.wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rcu_read_lock &obj_hash[i].lock irq_context: 0 &xa->xa_lock#20 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_NETLINK &rq->__lock irq_context: 0 sk_lock-AF_NETLINK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1922 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start rcu_node_0 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &knet->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx pool_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex fs_reclaim irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 (wq_completion)hci1#7 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &c->lock irq_context: 0 (wq_completion)bond0#15 irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1553 irq_context: 0 &dev->mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)bond0#29 irq_context: 0 (wq_completion)bond0#29 &rq->__lock irq_context: 0 cb_lock genl_mutex quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pcpu_lock stock_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1774 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 &mdev->req_queue_mutex quarantine_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1904 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1630 irq_context: 0 loop_validate_mutex loop_validate_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1630 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1630 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex.wait_lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem (&timer.timer) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock &x->wait#22 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem deferred_probe_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem device_links_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1630 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1629 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1629 irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1912 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1912 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1912 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1912 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1910 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1905 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1631 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1631 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1631 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1631 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1630 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1630 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy65 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#470 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#470 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#470 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#470 irq_context: 0 misc_mtx put_task_map-wait-type-override#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx put_task_map-wait-type-override#2 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1632 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1632 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1632 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy65 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1632 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1631 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1631 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1631 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &xt[i].mutex purge_vmap_area_lock quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1631 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1913 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1913 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1913 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &ep->mtx &____s->seqcount#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 namespace_sem &rq->__lock irq_context: 0 namespace_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1913 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1911 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1906 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy58 irq_context: 0 pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &data->open_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex uevent_sock_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#31 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#31 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#31 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &hdev->req_lock &____s->seqcount#2 irq_context: 0 &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#32 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#31 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#31 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#32 irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 irq_context: 0 (wq_completion)wg-crypt-wg1#16 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount irq_context: 0 rtnl_mutex &ul->lock#2 &c->lock irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem key irq_context: 0 rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 &hdev->req_lock &hdev->lock quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim &rcu_state.expedited_wq irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &sem->wait_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_node_0 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#60 irq_context: 0 nl_table_lock nl_table_wait.lock irq_context: 0 misc_mtx rfkill_global_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#60 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#33 irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#30 irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#33 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)bond0#34 irq_context: 0 rtnl_mutex team->team_lock_key#33 irq_context: 0 rtnl_mutex team->team_lock_key#33 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#33 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#33 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 lock irq_context: 0 rtnl_mutex team->team_lock_key#35 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &c->lock batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &c->lock kfence_freelist_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#63 irq_context: 0 (wq_completion)wg-kex-wg1#63 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) &rq->__lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 rtnl_mutex gdp_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex gdp_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 cb_lock rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (wq_completion)hci4#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#11 irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1696 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#353 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#353 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#353 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1984 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1984 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1980 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1980 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1980 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1975 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1698 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1698 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1698 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#519 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#519 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#519 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1698 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#519 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#517 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#517 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#517 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#516 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#356 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#356 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1697 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#356 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#354 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#354 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1697 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1985 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1985 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1985 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1985 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1981 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1976 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1699 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1699 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1698 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1698 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1698 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1698 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1698 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1698 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1986 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1986 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1986 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#520 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#520 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#520 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1986 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1982 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1977 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#520 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#518 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#517 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1700 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1700 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1700 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#357 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#357 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#355 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#355 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1700 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1699 irq_context: 0 pernet_ops_rwsem remove_cache_srcu &____s->seqcount irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#85 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#85 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#85 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1699 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#115 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#33 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 ebt_mutex &rq->__lock cpu_asid_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#33 &devlink_port->type_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#115 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#115 irq_context: 0 cb_lock key irq_context: 0 cb_lock pcpu_lock irq_context: 0 cb_lock percpu_counters_lock irq_context: 0 cb_lock pcpu_lock stock_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_owner irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle key#4 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem stock_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem key irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#9 key#13 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#115 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#121 &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex.wait_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[2] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#46 &n->list_lock irq_context: 0 kn->active#46 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 purge_vmap_area_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#394 irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 jbd2_handle irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &meta->lock irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 ebt_mutex key irq_context: 0 ebt_mutex pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle bit_wait_table + i irq_context: 0 &fc->uapi_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 ebt_mutex percpu_counters_lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work &p->pi_lock irq_context: 0 (wq_completion)events reg_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events reg_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#114 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1987 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1987 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1987 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1987 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1983 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1983 &rq->__lock irq_context: 0 rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1983 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1978 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1701 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1701 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1701 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1701 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1700 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock kfence_freelist_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &meta->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1700 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1988 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1988 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1988 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1988 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1984 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#521 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#521 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1934 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1934 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#330 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1929 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#330 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#330 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1652 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1652 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1652 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1979 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1702 irq_context: 0 &data->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1702 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1702 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#519 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#519 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#519 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1702 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1701 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#518 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1701 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1701 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1701 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1701 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex pool_lock#2 irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 key irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pcpu_lock stock_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#358 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#358 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#358 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#358 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#356 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#356 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1891 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1989 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1989 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1985 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1980 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1703 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1703 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1703 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1703 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1702 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1702 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1702 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1702 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1990 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1990 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1990 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1990 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1986 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1981 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1704 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1704 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1703 irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &c->lock irq_context: 0 &p->lock key irq_context: 0 (wq_completion)nfc4_nci_rx_wq#461 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#461 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1865 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1872 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy58 irq_context: 0 rtnl_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock key irq_context: 0 rtnl_mutex &idev->mc_lock pcpu_lock irq_context: 0 rtnl_mutex &idev->mc_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 kfence_freelist_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1703 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#522 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#522 &rq->__lock irq_context: 0 kn->active#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 kn->active#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#522 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#520 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#519 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1991 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1991 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1987 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1982 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1705 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1705 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1705 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1705 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1704 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1704 irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1704 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1704 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1992 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1992 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1992 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#523 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#523 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1992 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1988 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1988 &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex (&timer.timer) irq_context: 0 &dev->lock console_owner irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock &base->lock irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 &dum_hcd->dum->lock irq_context: 0 &x->wait#27 irq_context: 0 key#23 irq_context: softirq (&dum_hcd->timer) &x->wait#27 irq_context: softirq (&dum_hcd->timer) &x->wait#27 &p->pi_lock irq_context: softirq (&dum_hcd->timer) &x->wait#27 &p->pi_lock &rq->__lock irq_context: softirq (&dum_hcd->timer) &x->wait#27 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock pool_lock#2 irq_context: softirq (&dum_hcd->timer) &dum_hcd->dum->lock hcd_urb_list_lock irq_context: softirq (&dum_hcd->timer) lock#6 irq_context: softirq (&dum_hcd->timer) lock#6 kcov_remote_lock irq_context: softirq (&dum_hcd->timer) &x->wait#19 irq_context: softirq (&dum_hcd->timer) &x->wait#19 &p->pi_lock irq_context: softirq (&dum_hcd->timer) &x->wait#19 &p->pi_lock &rq->__lock irq_context: softirq (&dum_hcd->timer) &x->wait#19 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dum_hcd->timer) lock#6 &kcov->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex (&timer.timer) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1988 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#523 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#521 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#521 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex (&timer.timer) irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#520 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#359 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#359 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#359 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1983 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1707 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#523 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1999 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1999 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1999 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#527 irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#529 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#527 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#79 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dum_hcd->dum->lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#118 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#118 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &____s->seqcount irq_context: softirq rcu_read_lock hwsim_radio_lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex input_pool.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex req_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &x->wait#11 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1891 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1891 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1891 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1928 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1928 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#56 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#329 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#483 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#329 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#483 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#329 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#483 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#329 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#483 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#68 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#68 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#68 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#98 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#98 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#98 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#68 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_links_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1706 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1706 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1706 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex set_config_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex devtree_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1890 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1890 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1895 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1895 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#68 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dum_hcd->dum->lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 kn->active#28 &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock irq_context: 0 kn->active#28 &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#30 &dev->mutex &lock->wait_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex key irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex percpu_counters_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#30 &dev->mutex &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#30 &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1886 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1886 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1886 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1618 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1618 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1618 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1618 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1617 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1617 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1617 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1617 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#466 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#466 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#466 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#466 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#466 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#466 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1892 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1892 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1892 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1619 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1619 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1892 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1892 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1894 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1890 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1890 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1890 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1896 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1896 &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1896 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1896 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1622 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1621 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1900 irq_context: 0 &kcov->lock kcov_remote_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1899 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1899 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1900 &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#68 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#98 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#98 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#98 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#98 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#97 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#484 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#484 irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#484 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#484 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1938 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1938 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1929 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1938 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1938 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1936 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1931 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1931 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1931 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &kcov->lock kcov_remote_lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)hci4#11 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#359 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#357 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1706 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#357 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1705 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1939 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1937 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1932 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1654 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1654 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1653 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1653 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1933 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1655 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1655 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1654 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1654 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1656 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1655 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1655 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1655 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1655 irq_context: 0 &type->i_mutex_dir_key#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &xa->xa_lock#5 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#6 stock_lock irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock &wq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#99 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#99 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#485 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#485 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#99 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#485 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#485 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#485 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#485 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#98 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1705 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1705 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1705 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1705 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1705 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&local->tx_work) &rq->__lock irq_context: 0 (work_completion)(&local->tx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1993 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#98 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#98 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1993 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1989 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1984 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1984 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1984 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1707 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1707 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1707 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1707 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1706 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1706 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1706 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1706 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1994 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1994 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1994 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1994 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1990 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1990 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1990 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1985 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1708 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1708 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1707 irq_context: 0 &hdev->req_lock (wq_completion)hci0#12 irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1942 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#9 irq_context: 0 &hdev->req_lock (wq_completion)hci0#12 &rq->__lock irq_context: 0 &hdev->req_lock (wq_completion)hci0#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1995 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1995 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1995 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1995 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1991 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1986 irq_context: 0 (wq_completion)hci4#13 irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 &dev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex (&timer.timer) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_state_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_state_lock kernfs_notify_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dum_hcd->dum->lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sk->sk_timer) irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &base->lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock &rq->__lock irq_context: 0 &dev->mutex &udc->connect_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &udc->connect_lock udc_lock irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock pool_lock#2 irq_context: 0 &dev->mutex &udc->connect_lock udc_lock &queue->lock semaphore->lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &new_driver->dynids.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex device_links_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 snd_card_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &entry->access irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex proc_subdir_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex proc_subdir_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 pin_fs_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dum_hcd->dum->lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dum_hcd->dum->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dum_hcd->dum->lock &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 (&timer.timer) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_lock console_srcu console_owner irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 subsys mutex#71 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 subsys mutex#71 &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex req_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &x->wait#11 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex subsys mutex#71 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem snd_ctl_led_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 info_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 proc_subdir_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 proc_inum_ida.xa_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 proc_subdir_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex sound_loader_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &k->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex bus_type_sem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex dpm_list_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex req_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &x->wait#11 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex subsys mutex#71 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex subsys mutex#58 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex (usb_notifier_list).rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex subsys mutex#58 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 &kcov->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_state_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_state_lock kernfs_notify_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_state_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->files_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_loader_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 req_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 &type->i_mutex_dir_key/1 &xa->xa_lock#9 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &x->wait#11 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 subsys mutex#71 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->ctl_files_rwlock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key/1 &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &type->i_mutex_dir_key#2 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex subsys mutex#71 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &ent->pde_unload_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->power_sleep irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &card->files_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &entry->access irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &card->controls_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &card->controls_rwsem &xa->xa_lock#14 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex pin_fs_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &fsnotify_mark_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock mount_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex mount_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &x->wait#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd_urb_unlink_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kernfs_idr_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex (usb_notifier_list).rwsem usbfs_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex device_links_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 sb_writers &s->s_inode_list_lock irq_context: 0 sb_writers &obj_hash[i].lock irq_context: 0 sb_writers &sbinfo->stat_lock irq_context: 0 sb_writers &xa->xa_lock#9 irq_context: 0 sb_writers &fsnotify_mark_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &meta->lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex stock_lock irq_context: 0 udc_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 udc_lock rcu_read_lock &rq->__lock irq_context: 0 udc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock cpu_asid_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &base->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#121 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#122 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1602 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1602 &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#35 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#33 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1668 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1668 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1668 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1667 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1667 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1667 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1667 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1667 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1667 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#337 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#337 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#337 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#337 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#336 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#336 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#336 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#336 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#103 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#103 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#103 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#72 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#72 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1226 irq_context: 0 uevent_sock_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1366 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1366 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1369 irq_context: 0 (wq_completion)events_unbound &rq->__lock irq_context: 0 (wq_completion)events_unbound &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_node_0 irq_context: 0 sb_writers#4 &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1373 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#122 &rq->__lock irq_context: softirq (&dum_hcd->timer) &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1377 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1374 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1375 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1375 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1375 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1375 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1376 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#347 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#347 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#248 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#248 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1376 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1376 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1377 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1377 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1377 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1377 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex batched_entropy_u8.lock irq_context: 0 nfc_devlist_mutex kfence_freelist_lock irq_context: 0 nfc_devlist_mutex &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1378 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1378 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1378 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#348 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#348 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#348 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#348 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#348 &rq->__lock irq_context: 0 &u->iolock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#348 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#249 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#37 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1379 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1379 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1379 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1379 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1380 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#37 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#38 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1380 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1380 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1381 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#37 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 udc_lock console_owner_lock irq_context: 0 udc_lock console_owner irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#38 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#19 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1381 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1381 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx remove_cache_srcu &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1382 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1382 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1383 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1383 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1383 &rq->__lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1383 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1242 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1242 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1242 irq_context: 0 &xt[i].mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1383 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1383 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1384 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1384 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1384 &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1244 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1384 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1384 irq_context: 0 &knet->mutex &rq->__lock irq_context: 0 &knet->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1385 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem fs_reclaim irq_context: 0 purge_vmap_area_lock &meta->lock irq_context: 0 purge_vmap_area_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1390 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1390 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount#2 irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock batched_entropy_u8.lock irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex &udc->connect_lock &queue->lock kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1391 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1391 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1392 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1392 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1392 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1392 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1392 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1394 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1253 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1253 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1253 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1253 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1395 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1395 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1395 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1395 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1254 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1254 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1396 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1396 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1396 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1396 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1397 irq_context: softirq init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1397 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1397 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1398 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1398 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1399 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1399 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1399 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1257 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1257 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1257 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1399 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1400 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1400 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1400 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1258 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1258 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1258 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1401 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1401 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1259 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1401 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1401 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1417 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1421 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1421 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1424 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1602 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1603 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1603 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1451 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1451 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1451 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1452 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1603 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1603 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1604 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#30 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1433 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1433 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#15 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1433 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1433 &rq->__lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 krc.lock &base->lock irq_context: 0 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#29 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#30 irq_context: 0 rtnl_mutex dev_addr_sem batched_entropy_u8.lock irq_context: 0 rtnl_mutex dev_addr_sem kfence_freelist_lock irq_context: 0 fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1436 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1436 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1436 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1436 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1437 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1437 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1437 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1440 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1440 &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#5 &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#5 pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work stock_lock irq_context: 0 &dev->dev_mutex &n->list_lock irq_context: 0 &dev->dev_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1297 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1604 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1442 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1442 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1444 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1446 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1449 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1449 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1450 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1325 irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1469 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1469 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1470 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1470 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1326 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1326 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1470 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1470 &rq->__lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &rq->__lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1470 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1327 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1471 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1471 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1328 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1328 &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1473 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1476 &rq->__lock irq_context: 0 sb_writers#4 &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1476 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1477 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1477 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1477 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1477 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1477 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1333 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1333 &rq->__lock irq_context: 0 btf_idr_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1478 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1478 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1477 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1479 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1479 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1335 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1335 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1335 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1479 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1479 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1480 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1480 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1481 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1481 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1337 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1337 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: softirq (&pool->idle_timer) &pool->lock irq_context: softirq (&pool->idle_timer) &pool->lock &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1338 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1482 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1482 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1483 irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1483 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1484 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1484 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1484 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1484 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1485 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1485 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#258 irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1485 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1485 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1485 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1487 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1487 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1488 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1488 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1488 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1488 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1489 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1345 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1489 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1490 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1490 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1346 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex console_owner irq_context: softirq (&peer->timer_persistent_keepalive) kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1346 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1346 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1346 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1490 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1490 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1490 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1490 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1491 irq_context: 0 &u->iolock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &u->iolock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &u->iolock &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1491 irq_context: 0 &u->iolock &mm->mmap_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1347 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1347 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#33 &dev->mutex &lock->wait_lock irq_context: 0 &p->lock &of->mutex kn->active#33 &dev->mutex &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#33 &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1491 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1491 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1492 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1492 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1348 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1348 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1348 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1492 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1492 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1493 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1493 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1349 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1356 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1501 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1501 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#122 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#122 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#123 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#123 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#124 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#124 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1502 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1502 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1502 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1502 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1502 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1357 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1358 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1502 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1502 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1503 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1503 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1358 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1358 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1358 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &meta->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1359 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1503 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1503 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1503 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1503 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1504 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1504 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1505 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1505 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1505 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1505 &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1505 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1506 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1506 irq_context: 0 tty_mutex stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1362 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1506 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1362 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1506 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1362 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1362 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1507 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1507 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1508 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1508 irq_context: 0 tty_mutex &rq->__lock irq_context: 0 tty_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex tty_mutex.wait_lock irq_context: 0 tty_mutex.wait_lock irq_context: 0 tty_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1512 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1513 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1513 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1368 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1368 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1368 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1369 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#48 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1513 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1513 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1513 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1513 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1514 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#33 &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1514 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1369 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1369 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1369 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1514 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1514 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1515 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1515 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#363 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1517 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1517 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1371 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#364 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#364 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#364 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1518 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1518 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1518 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1519 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1519 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1519 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1519 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1520 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1520 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1373 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1373 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1373 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1373 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1373 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1374 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1374 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &list->lock#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1520 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1520 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1520 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1520 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1522 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#394 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1730 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1730 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1730 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1511 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1511 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1511 irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1511 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#49 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1514 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#278 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#29 &rq->__lock irq_context: 0 kn->active#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#278 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#278 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#49 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#63 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#18 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1731 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1731 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#64 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#64 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#397 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#397 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1731 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1452 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1452 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1452 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1453 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1615 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1616 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1616 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1616 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1616 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#125 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#125 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#126 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#126 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1616 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1616 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1463 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1463 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1463 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1463 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc6_nci_rx_wq#52 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#52 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ep->mtx pcpu_lock stock_lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1915 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1915 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1913 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1908 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#268 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#5 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1464 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1464 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1465 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1617 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1618 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 rtnl_mutex dev_hotplug_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_hotplug_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_hotplug_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#73 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 &iopt->domains_rwsem &rq->__lock irq_context: 0 &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rnp->exp_wq[2] irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#73 rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1914 irq_context: 0 &data->open_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &iopt->domains_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1914 &rq->__lock irq_context: 0 (wq_completion)hci3#7 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1914 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1914 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 &type->s_umount_key#21/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1912 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1912 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1912 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1907 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1907 &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &n->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#3 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1618 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1465 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1465 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1466 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1619 irq_context: 0 &mm->mmap_lock fs_reclaim rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1622 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1622 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#374 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1624 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1625 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1628 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1628 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1628 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1625 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1472 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1636 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1635 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1635 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#472 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#472 irq_context: 0 &type->i_mutex_dir_key/1 fs_reclaim irq_context: 0 &type->i_mutex_dir_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1918 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1918 &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1918 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 fs_reclaim irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1918 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1922 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1920 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1915 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1640 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1640 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1639 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1626 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1639 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#476 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#476 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1634 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1634 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1634 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1634 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1633 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1633 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#117 irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2002 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1626 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1626 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 key irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock x25_neigh_list_lock irq_context: softirq rcu_read_lock x25_list_lock irq_context: softirq rcu_read_lock x25_forward_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#24 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1628 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#23 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#23 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#45 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#46 irq_context: 0 (wq_completion)wg-kex-wg0#49 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &(&fn_net->fib_chain)->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &xa->xa_lock#20 &n->list_lock irq_context: 0 &xa->xa_lock#20 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock semaphore->lock#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1629 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1629 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &dum_hcd->dum->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &x->wait#19 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &base->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 hcd->bandwidth_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1633 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1554 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1554 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#10 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1409 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1409 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1558 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1558 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1558 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1559 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1559 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#18 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1512 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1512 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1512 irq_context: 0 kn->active#12 &n->list_lock irq_context: 0 kn->active#12 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1512 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#396 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1512 &rq->__lock irq_context: 0 &p->lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock fs_reclaim &obj_hash[i].lock irq_context: 0 &p->lock fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#396 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#396 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#396 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#396 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1512 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1629 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#396 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1512 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mdev->req_queue_mutex &mdev->graph_mutex &rq->__lock irq_context: 0 &mdev->req_queue_mutex &mdev->graph_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#63 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#63 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#63 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#63 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1629 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1630 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1630 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1630 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1630 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1507 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1507 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1507 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1507 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1507 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1507 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1507 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1507 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#60 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#60 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#60 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#60 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#48 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#48 irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#60 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#48 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#48 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#275 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#275 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#275 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#275 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1731 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1731 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#275 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#275 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#275 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#275 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1731 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#390 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex hcd->bandwidth_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#390 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#390 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#390 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#390 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1727 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1727 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1727 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1727 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1726 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1726 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1726 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#45 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#45 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#45 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#45 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#45 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#45 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#272 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#272 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#272 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#272 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#272 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#272 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#272 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#272 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#57 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#57 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#57 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#57 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#49 &n->list_lock irq_context: 0 kn->active#49 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#47 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1630 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#47 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#48 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1630 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1508 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1508 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1508 irq_context: 0 &ep->mtx &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &lock->wait_lock irq_context: 0 &p->lock &of->mutex kn->active#30 &lock->wait_lock irq_context: 0 &p->lock &of->mutex kn->active#30 &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pool_lock#2 irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 &p->lock &of->mutex kn->active#30 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#30 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#30 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kn->active#30 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kn->active#30 &root->deactivate_waitq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kn->active#30 sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kn->active#30 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex kn->active#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &k->k_lock irq_context: 0 &ep->mtx &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &ep->mtx &mm->mmap_lock pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1508 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#61 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#61 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1#11 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1509 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1509 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1509 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1509 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1509 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#394 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#394 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#394 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx (console_sem).lock irq_context: 0 misc_mtx &wq->mutex irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#5 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#5 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1631 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1631 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1631 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1631 irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1632 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1633 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1634 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1634 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock semaphore->lock#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1636 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1636 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1637 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &pipe->mutex/1 &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond0#18 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1639 irq_context: 0 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1730 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1730 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1730 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1730 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1729 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#57 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#57 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#57 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1640 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1640 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#5 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#12 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1661 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1661 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1661 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1662 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1662 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &base->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &base->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &base->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1662 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1662 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1662 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1662 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1663 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1663 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1663 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1664 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1664 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1664 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1664 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1665 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1665 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#27 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#27 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#53 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#54 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1730 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1728 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1728 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#26 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#51 irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1670 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1670 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1670 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1670 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1671 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1671 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1671 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) quarantine_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1728 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#51 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#51 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex leds_list_lock &rq->__lock irq_context: 0 cb_lock genl_mutex leds_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#524 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#524 irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#522 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#522 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#522 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#72 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#72 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#103 &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#103 &obj_hash[i].lock irq_context: 0 &dev->mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1952 irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tomoyo_ss (kmod_concurrent_max).lock irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1643 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1654 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1659 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1659 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1659 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1659 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1660 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1730 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#127 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#127 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1660 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1660 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1660 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1660 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 kn->active#53 &n->list_lock irq_context: 0 kn->active#53 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock quarantine_lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#9 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1916 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1916 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1916 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1911 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#476 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1505 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#62 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &n->list_lock &c->lock irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#52 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1666 irq_context: 0 (wq_completion)bond0#26 irq_context: 0 &p->lock &of->mutex &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex pool_lock#2 irq_context: 0 (wq_completion)bond0#26 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1668 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 &group->mark_mutex batched_entropy_u8.lock irq_context: 0 &group->mark_mutex kfence_freelist_lock irq_context: 0 &hdev->req_lock (wq_completion)hci0#8 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) quarantine_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#2 rcu_node_0 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &meta->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#21 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy50 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#52 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1674 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1674 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1675 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#22 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#22 irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)nfc39_nci_rx_wq irq_context: 0 (wq_completion)nfc39_nci_tx_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc47_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc46_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy34 irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &obj_hash[i].lock irq_context: 0 lock pidmap_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1683 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock lock#4 irq_context: 0 sb_writers#3 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1685 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1685 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1685 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 &ei->xattr_sem irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#476 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1689 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#324 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1689 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1687 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1689 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1687 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1687 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#31 irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1727 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1694 irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 &ei->i_data_sem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 (wq_completion)netns net_cleanup_work quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1698 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1698 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx console_owner_lock irq_context: 0 misc_mtx console_owner irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1700 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1700 &rq->__lock irq_context: 0 lock map_idr_lock &n->list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#542 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1702 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1702 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1702 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1701 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1487 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex minors_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1708 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle kfence_freelist_lock irq_context: 0 rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1709 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 binderfs_minors_mutex irq_context: 0 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &sb->s_type->i_mutex_key#18 irq_context: 0 &sb->s_type->i_mutex_key#18 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#18 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock key irq_context: 0 &sighand->siglock batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1710 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &uhid->devlock &n->list_lock irq_context: 0 &uhid->devlock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem bit_wait_table + i irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 nf_sockopt_mutex stock_lock irq_context: 0 nf_sockopt_mutex key irq_context: 0 &sb->s_type->i_mutex_key#10 &knet->mutex &rq->__lock irq_context: 0 nf_sockopt_mutex pcpu_lock irq_context: 0 nf_sockopt_mutex percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1710 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1712 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1712 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1712 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1713 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1713 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1714 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1715 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1715 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1716 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1716 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1716 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1717 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1717 &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &queue->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET key irq_context: 0 sk_lock-AF_INET pcpu_lock irq_context: 0 &hdev->req_lock (wq_completion)hci4#8 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#52 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#52 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#52 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#41 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#41 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#41 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#69 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#41 irq_context: 0 (wq_completion)wg-kex-wg2#69 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#41 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#41 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#386 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#386 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#386 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#386 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#386 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#386 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1501 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1501 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1501 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1501 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1501 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1501 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1783 irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1776 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1422 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1501 &rq->__lock irq_context: 0 sk_lock-AF_INET percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1422 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1279 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1776 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#278 &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#278 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#278 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#278 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1515 irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#398 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#7 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#398 irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#398 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#398 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#280 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1733 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1516 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1517 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1779 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1779 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1775 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1775 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1775 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1780 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1780 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1780 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1780 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1780 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1776 irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1781 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1781 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1781 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1781 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1781 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1781 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1781 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 sk_lock-AF_INET pcpu_lock stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#283 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#283 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#283 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#283 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#403 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#403 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1520 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1520 &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1523 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#408 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#408 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#408 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#408 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#408 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1741 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1741 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem key irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1744 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1525 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1525 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1525 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1525 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1745 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1529 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1530 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1530 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev_instance->mutex remove_cache_srcu pool_lock#2 irq_context: 0 &dev_instance->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &dev_instance->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &queue->lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_internal &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#54 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#54 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#54 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#540 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#537 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#537 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#73 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 lweventlist_lock &dir->lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#36 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#74 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#369 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#369 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#369 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#369 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#367 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#367 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#367 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &xa->xa_lock#20 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#121 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#121 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#121 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#121 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#121 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy53 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#26 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#52 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#51 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#26 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#51 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1777 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#52 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#324 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#323 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#323 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#32 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#121 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#121 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#120 irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#27 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg0#55 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1761 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1761 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1761 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1761 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock stock_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 remove_cache_srcu &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1776 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1776 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1763 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1763 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1763 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1763 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1763 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1760 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1541 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1541 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1542 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1542 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1542 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1543 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1543 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1543 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1545 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1545 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_IPGRE irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work stock_lock irq_context: 0 kn->active#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &meta->lock irq_context: 0 pernet_ops_rwsem sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1773 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1773 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: softirq slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq slock-AF_INET#2 fill_pool_map-wait-type-override &c->lock irq_context: softirq slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &mdev->req_queue_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1776 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1777 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1777 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1777 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1777 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1777 irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1778 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1778 &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &p->lock &of->mutex kn->active#12 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#50 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#51 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1279 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#29 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1784 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy51 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1785 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1785 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy51 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 irq_context: 0 (wq_completion)wg-crypt-wg2#16 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1940 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1935 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#100 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#100 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#100 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#39 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#20 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &pcp->lock &zone->lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1806 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1806 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1560 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#293 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#429 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#429 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#429 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1813 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1813 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1813 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1815 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1815 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1815 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1819 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#88 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#97 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1826 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1826 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1826 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1826 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1828 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_node_0 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#103 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1951 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1951 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#72 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#103 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#102 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#102 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1947 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1942 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#102 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#102 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1952 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1952 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1952 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1948 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1657 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1657 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1657 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1657 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1656 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1656 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1656 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1656 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1656 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1656 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#100 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#100 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#99 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#331 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#99 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#331 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#330 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#330 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#69 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#69 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#69 &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#69 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#30 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#30 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#30 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#28 irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1943 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1943 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1943 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1941 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1658 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1658 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1658 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1936 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1936 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1936 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1657 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1657 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1657 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1657 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1657 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1657 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#486 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#486 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#486 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#486 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#486 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#486 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1944 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1944 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1942 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1937 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1937 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#487 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#487 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#1659 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#1659 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#1659 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#487 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#1659 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#1659 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#1659 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#487 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#487 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1660 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1945 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1945 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1945 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1946 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#332 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#332 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#332 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#331 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#331 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1947 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1947 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1943 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1938 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1661 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1661 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1661 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1660 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1660 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1660 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1660 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1660 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1660 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1662 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1662 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1661 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1661 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1661 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1661 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#333 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#332 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#332 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#489 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#489 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#489 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#489 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#489 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#489 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1663 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1663 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1662 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1662 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1662 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#490 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#490 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#490 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#490 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#490 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#490 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#490 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1948 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1948 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1948 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1948 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1944 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1939 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#491 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#491 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#491 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#491 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#491 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#491 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#31 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#31 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#31 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#492 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#492 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#31 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#31 irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#29 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#29 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#492 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#70 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#70 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#70 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &c->lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#70 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1664 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1664 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1663 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &n->list_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 kn->active#50 remove_cache_srcu irq_context: 0 kn->active#50 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#50 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 kn->active#50 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 &type->s_umount_key#47 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex.wait_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex nf_hook_mutex &n->list_lock irq_context: 0 rtnl_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1943 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#22 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2007 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2007 &rq->__lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu irq_context: 0 kn->active#12 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy64 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#120 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#120 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&hcd->wakeup_work) &dev->mutex &hub->status_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy63 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock quarantine_lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#36 &devlink_port->type_lock irq_context: 0 &hdev->req_lock &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#29 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#58 irq_context: 0 rtnl_mutex dev_addr_sem &base->lock irq_context: 0 rtnl_mutex dev_addr_sem &base->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1567 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1417 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1417 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1417 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1417 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1417 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1418 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1568 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1568 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1418 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1418 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1418 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1419 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1419 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1419 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1419 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#299 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#299 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#299 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#299 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#128 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#128 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#128 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#129 irq_context: 0 sb_writers#5 quarantine_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#9 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#129 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#129 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#129 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#130 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#130 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#130 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#130 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#131 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#131 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#131 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#131 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#132 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#132 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#132 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#132 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1420 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1569 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#133 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#133 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#134 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#134 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#134 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1569 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1570 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1570 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1420 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1421 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1421 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 console_owner irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1421 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1421 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1422 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1422 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1422 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1571 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1572 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1422 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1663 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1663 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1664 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 gdp_mutex &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex uevent_sock_mutex rcu_node_0 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1422 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1422 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1423 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1572 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1573 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1573 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1573 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1573 &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#26 irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#135 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#135 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1423 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1423 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1423 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1423 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1423 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1424 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#136 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#136 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#137 irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#47 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1573 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1573 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1574 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1574 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1424 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1424 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1424 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1425 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#370 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#370 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#370 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#370 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1574 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1574 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1574 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1574 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1575 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1575 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1425 irq_context: 0 &dev->mutex &rcu_state.expedited_wq irq_context: 0 &dev->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#137 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#138 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#138 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1845 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1426 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1426 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#496 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#496 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#496 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#496 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#496 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#496 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#497 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#497 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1426 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1426 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1426 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1427 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1427 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1427 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1427 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1427 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1427 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1427 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1428 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1579 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1579 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1428 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1428 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1428 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1429 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1580 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1580 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1430 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1581 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1430 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1430 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1430 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1430 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1431 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1581 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1581 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1581 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1581 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1582 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1582 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1582 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1431 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1431 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1431 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1432 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1582 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1582 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1583 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1583 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1583 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1583 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1432 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1432 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1432 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1432 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1433 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1433 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1433 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1433 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1433 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1434 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1434 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1434 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1434 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1434 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1435 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1435 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex input_pool.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1907 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#5 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETROM irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1962 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#505 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#505 &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &meta->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &base->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#13 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 &u->iolock &base->lock irq_context: 0 &u->iolock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#14 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&conn->pending_rx_work) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &meta->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#521 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1709 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1709 irq_context: 0 sb_writers#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1708 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem stock_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#24 &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#24 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#24 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1708 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem key irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1708 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1708 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#360 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#360 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#360 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#360 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#358 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 cb_lock genl_mutex rtnl_mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem nl_table_wait.lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#358 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#358 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#358 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#358 &rq->__lock irq_context: 0 vmap_purge_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1633 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1633 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1845 irq_context: 0 put_task_map-wait-type-override#3 &rq->__lock irq_context: 0 put_task_map-wait-type-override#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#306 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[1] irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#306 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1846 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#446 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1591 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#446 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#446 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1632 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1632 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#114 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#306 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#115 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#61 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1847 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy57 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#85 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1805 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1593 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1593 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#117 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1593 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1595 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1595 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1594 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1596 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#119 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1850 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1850 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1850 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#120 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1599 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1599 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1599 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1599 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1598 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1598 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#452 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#452 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 sb_writers#7 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#452 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#452 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#452 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#121 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#453 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1852 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &meta->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1601 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1601 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#122 irq_context: 0 misc_mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1601 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#25 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#25 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#25 irq_context: 0 misc_mtx &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#124 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#126 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#455 irq_context: 0 kn->active#4 fs_reclaim &cfs_rq->removed.lock irq_context: 0 kn->active#4 fs_reclaim &obj_hash[i].lock irq_context: 0 kn->active#4 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#129 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1604 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1604 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1603 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#133 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#458 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#458 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1859 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1859 &rq->__lock irq_context: 0 &ndev->req_lock &wq->mutex rcu_node_0 irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1859 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1608 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock &hub->status_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1862 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1862 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1862 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1858 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1863 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1865 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1865 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[3] irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1866 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1866 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1868 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1868 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#92 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex stock_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1878 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1878 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1878 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1279 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 kfence_freelist_lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1841 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1591 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1591 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#445 irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1884 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1884 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1884 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1884 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#139 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#139 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1646 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1646 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1931 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1931 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1931 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1931 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1929 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1924 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1648 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1648 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rose_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rlock-AF_ROSE irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1932 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1932 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1925 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1649 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1649 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#18 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex sysctl_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2006 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#32 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#32 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#36 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#36 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#18 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy56 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy56 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy56 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1886 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1886 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy55 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1619 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1619 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1892 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1805 irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#24 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#48 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)bond0#21 irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2006 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2002 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#42 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1923 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1637 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1637 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1636 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1636 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1633 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1636 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#364 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2007 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1636 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#473 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex key irq_context: 0 (wq_completion)wg-kex-wg1#52 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &pipe->mutex/1 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1884 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1671 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy49 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1672 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy49 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_state.exp_mutex pcpu_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex percpu_counters_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2010 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_owner_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex console_owner irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1726 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1723 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->ipv4.ra_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &hashinfo->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#473 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#473 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#473 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &mm->mmap_lock rcu_node_0 irq_context: 0 rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 rtnl_mutex fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#473 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#473 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 &dev->mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 &dev->mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#29 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#57 irq_context: 0 &xt[i].mutex free_vmap_area_lock quarantine_lock irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1553 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#58 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#29 irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_state.exp_wake_mutex.wait_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#58 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex sched_map-wait-type-override &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1404 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 pcpu_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1628 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#370 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2019 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2019 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2019 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2015 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#83 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1404 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1619 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#445 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1466 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1466 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1466 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#445 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#445 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1467 irq_context: 0 misc_mtx (wq_completion)nfc9_nci_cmd_wq#23 irq_context: 0 misc_mtx (wq_completion)nfc9_nci_cmd_wq#23 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc9_nci_cmd_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#497 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#497 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#498 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#498 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#498 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#499 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#499 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1953 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1953 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1953 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1953 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1949 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1949 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1949 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#338 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#338 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#338 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1944 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1944 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1944 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &obj_hash[i].lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 lock prog_idr_lock &n->list_lock irq_context: 0 lock prog_idr_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1365 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1365 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1365 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1366 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1366 irq_context: 0 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1366 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1367 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1367 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy39 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1367 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1367 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1367 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1368 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1368 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1368 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1368 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1369 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1228 irq_context: 0 sb_writers#4 oom_adj_mutex oom_adj_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1370 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1370 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1370 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1402 irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1402 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1260 irq_context: 0 sb_writers#4 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1260 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1261 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1402 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1402 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1403 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1403 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1261 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1261 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1261 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1261 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1262 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1403 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1403 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1403 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1404 irq_context: 0 kn->active#13 &____s->seqcount#2 irq_context: 0 kn->active#13 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#13 &rq->__lock irq_context: 0 kn->active#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex.wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1404 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#67 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1404 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1264 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1264 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1264 irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#353 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1405 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1405 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1406 irq_context: 0 &hdev->req_lock (wq_completion)hci1#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1406 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1407 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1407 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1407 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1407 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#354 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1409 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1409 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1410 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1410 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle key#4 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1280 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx sched_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 &p->lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1411 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1422 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1412 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1412 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1422 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1423 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1423 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1423 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1423 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond0#22 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1280 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1280 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#17 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex key#22 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1273 irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1415 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1416 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1416 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1416 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1416 irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &cfs_rq->removed.lock irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1416 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1416 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1416 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1416 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1417 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1417 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1274 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1274 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1274 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1417 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1418 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1418 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1418 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1418 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1275 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1275 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1275 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1923 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1276 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1280 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1280 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1418 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1280 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy31 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1281 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1281 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1281 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1419 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 pool_lock#2 irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 key irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy31 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1423 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1423 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1423 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1423 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1424 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1424 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1424 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1424 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1424 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1281 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1281 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1281 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1281 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1281 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1424 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1424 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1425 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1425 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1425 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1425 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1426 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1426 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1289 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1289 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &base->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1289 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1290 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1428 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1428 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1428 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1429 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1619 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1620 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1620 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1467 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1467 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1467 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1440 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1297 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1297 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1298 irq_context: 0 &type->s_umount_key#21/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1441 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1441 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1442 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1442 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1442 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1443 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1443 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1299 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback stock_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1443 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1443 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1443 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1443 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1444 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1444 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1444 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1445 irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1445 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1301 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1301 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1301 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1302 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1445 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1445 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1445 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1445 irq_context: 0 &root->kernfs_iattr_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &root->kernfs_iattr_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1446 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1446 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1446 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1302 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1446 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1446 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1447 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1447 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1447 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1447 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1303 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1303 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1303 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1303 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1447 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1447 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1448 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 remove_cache_srcu rcu_node_0 irq_context: 0 kn->active#4 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 kn->active#4 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#4 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1884 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1880 irq_context: 0 (wq_completion)wg-kex-wg0#61 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#61 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 misc_mtx pcpu_alloc_mutex rcu_node_0 irq_context: 0 misc_mtx pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1923 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1923 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1921 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1921 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1917 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1917 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1448 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1448 &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem &hub->status_mutex &meta->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex ehci_cf_port_reset_rwsem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1304 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1448 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1448 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1448 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1449 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1449 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1450 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1306 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1306 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1306 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1307 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1307 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1307 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1308 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#358 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#358 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#358 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1451 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1451 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1805 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1505 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#64 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#64 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#397 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#397 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#50 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#50 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#50 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#50 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#397 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#50 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1505 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1505 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1505 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1505 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#58 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#58 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#58 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#58 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#58 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#397 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#50 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#50 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#19 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#277 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#277 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1732 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1732 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1732 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1732 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1732 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1732 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1732 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1513 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#62 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#62 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1513 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1513 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1451 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1452 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1452 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1452 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1452 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1452 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1308 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1452 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1452 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1452 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1309 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1453 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1454 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1454 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1454 irq_context: 0 &dev->mutex leds_list_lock &rq->__lock irq_context: 0 &dev->mutex leds_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1310 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1310 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1310 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1310 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1454 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1454 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1454 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1455 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1455 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1311 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1455 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1455 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1456 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1456 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1456 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1457 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1457 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1313 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1313 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1313 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback stock_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1314 irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1457 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1457 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1458 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1458 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1314 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1314 &rq->__lock irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 &n->list_lock irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1458 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1458 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1458 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1458 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1315 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1459 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#277 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1513 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#277 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1513 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1513 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1513 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1513 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#46 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#46 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#46 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#46 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#46 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1728 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#16 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1728 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1727 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#9 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1506 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1506 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1506 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1506 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1506 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#391 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#391 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#391 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#391 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#391 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#391 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#278 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1631 pool_lock#2 irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 misc_mtx &wq->mutex &pool->lock irq_context: 0 misc_mtx &wq->mutex &x->wait#10 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock &rq->__lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#22 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#44 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#43 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1467 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1467 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1468 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1468 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1468 irq_context: 0 kn->active#14 &n->list_lock irq_context: 0 kn->active#14 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#22 irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1554 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1554 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1555 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1631 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1631 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1631 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx wq_mayday_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1921 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1916 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1641 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1641 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1640 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1640 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1640 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1640 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#477 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#477 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#477 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#477 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#477 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1916 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1555 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1916 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1916 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1555 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1916 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1914 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1914 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1914 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1406 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1406 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#43 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1406 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1407 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#44 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1555 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1555 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1909 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1909 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1909 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1635 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1635 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1635 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1635 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1634 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1634 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1634 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1634 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#471 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#471 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#471 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#471 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#471 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#471 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#471 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1630 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx key irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1917 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#477 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#477 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#325 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#325 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#325 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#324 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#324 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#324 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#64 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1917 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1917 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1917 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1917 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1922 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1922 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1922 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1919 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1919 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1919 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1917 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1912 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1912 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1912 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1638 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1638 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1637 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1637 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#474 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#474 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#474 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#474 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#474 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#474 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#323 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#323 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#323 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#323 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#322 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#322 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1920 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1920 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1920 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1920 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1920 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1920 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1918 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1913 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1921 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1921 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1921 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1921 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1921 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1921 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1919 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1914 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1639 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1639 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1638 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1638 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#475 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#475 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#475 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#475 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1924 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#32 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#63 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#64 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#32 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1642 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1642 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1925 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1925 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1925 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1642 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1925 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1923 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1918 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1641 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1641 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1641 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1641 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#478 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#478 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#478 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#478 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#478 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#326 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#326 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#325 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#325 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#325 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#478 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#325 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#94 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#94 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#94 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#94 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#94 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#94 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#327 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#327 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#326 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#326 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc5_nci_tx_wq#326 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#326 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#479 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#479 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1643 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1643 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#56 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1642 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1642 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1926 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1926 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1919 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#96 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#96 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#96 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#96 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#96 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#96 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#96 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#95 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#328 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#328 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#327 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#327 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#327 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#327 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1927 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1925 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1920 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1555 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1556 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1556 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#22 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1556 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1556 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1407 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#43 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1920 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1920 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1644 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1644 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#28 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1644 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1643 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1643 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1643 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#55 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1926 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1921 irq_context: 0 rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1645 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1645 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1644 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1644 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1644 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1929 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1929 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1929 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1927 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1922 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1646 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1646 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1645 irq_context: 0 pcpu_alloc_mutex rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1645 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1645 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1645 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#28 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1930 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#44 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1928 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1928 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1647 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1647 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1647 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1647 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1646 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1646 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#22 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1557 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1557 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1557 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1557 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1558 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1646 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1646 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1648 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1648 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1647 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1647 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1647 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1647 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1648 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1648 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1648 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1648 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1801 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#480 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#480 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#480 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#480 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#480 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#480 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#32 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#32 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#3 irq_context: 0 cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#8 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#3 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#3 &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_tx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#3 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#3 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_rx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#9 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#338 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#337 irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1669 irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#18 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1669 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1559 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1559 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1560 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1561 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1669 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#337 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#337 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#337 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1669 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1668 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1668 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1668 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1668 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#104 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#104 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#104 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#104 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#104 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#4 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#73 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#73 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#73 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1954 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1954 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1954 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#103 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#103 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#103 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#103 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#103 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1954 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#73 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#73 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1950 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1950 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1950 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1945 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#73 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#36 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex key irq_context: 0 cb_lock genl_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex percpu_counters_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx percpu_counters_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1632 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1632 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1632 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1632 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1632 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1633 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1633 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1469 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 nl_table_lock &c->lock irq_context: 0 tomoyo_ss rcu_read_lock stock_lock irq_context: 0 tomoyo_ss rcu_read_lock key irq_context: 0 &iopt->domains_rwsem irq_context: 0 nfc_devlist_mutex rcu_node_0 irq_context: 0 nfc_devlist_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock percpu_counters_lock irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1634 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1634 &rq->__lock irq_context: 0 rtnl_mutex wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 &iopt->domains_rwsem &iopt->iova_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1474 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1474 irq_context: 0 cb_lock rtnl_mutex stock_lock irq_context: 0 cb_lock rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex key irq_context: 0 cb_lock rtnl_mutex pcpu_lock irq_context: 0 cb_lock rtnl_mutex percpu_counters_lock irq_context: 0 cb_lock rtnl_mutex pcpu_lock stock_lock irq_context: 0 cb_lock rtnl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1634 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex pool_lock#2 irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1634 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1635 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1635 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1635 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1635 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1635 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1635 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1635 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1636 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1501 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &pcp->lock &zone->lock irq_context: 0 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#13 irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#3 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1723 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1723 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#269 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#13 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#13 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1724 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers#3 sb_internal &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#388 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#388 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#388 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#56 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#56 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#56 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#56 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#56 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#56 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1504 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1504 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#15 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1782 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1782 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1782 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1782 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1782 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1782 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1782 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#70 irq_context: 0 &hdev->req_lock (wq_completion)hci1#14 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1774 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1774 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1774 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1774 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1770 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#389 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#389 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1725 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#5 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_rx_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#6 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1907 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1907 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1907 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1905 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1900 irq_context: 0 &hdev->req_lock &hdev->lock pool_lock irq_context: 0 sb_writers#9 remove_cache_srcu irq_context: 0 sb_writers#9 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#9 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#9 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#9 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1778 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1778 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1778 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1774 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1774 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1774 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1778 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1778 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1778 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy52 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1783 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1783 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1783 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1783 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1779 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1784 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1784 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1784 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1636 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1636 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1636 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1636 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1636 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1637 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1637 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1637 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1637 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1637 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1638 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1638 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1638 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1638 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1638 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1638 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1638 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1638 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1639 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1639 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1639 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1640 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1640 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#9 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1640 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1641 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1641 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1641 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1641 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1641 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1641 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1642 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1642 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1642 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1642 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1642 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1642 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1643 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1643 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1643 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1643 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1643 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1643 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1644 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1644 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1644 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1644 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1644 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1644 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1645 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1645 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1645 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1645 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1645 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1723 irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 tomoyo_ss &x->wait#17 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1645 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1645 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1645 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1645 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1645 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1646 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1646 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1646 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1647 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1647 irq_context: 0 tomoyo_ss running_helpers_waitq.lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#36 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#73 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#73 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#73 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#34 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1647 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1647 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1648 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1648 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1649 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1649 &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#7 irq_context: 0 &f->f_pos_lock sb_writers#7 fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1650 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1650 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1650 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1651 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1651 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1652 irq_context: 0 kn->active#53 remove_cache_srcu irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1653 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1653 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1654 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1655 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1655 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1655 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1655 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1656 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1656 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1656 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1656 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1656 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1656 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1657 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#34 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#24 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#24 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1657 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1657 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1657 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#52 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1675 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1675 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1675 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#51 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1480 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1480 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2019 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2019 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2015 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1481 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1481 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1481 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1481 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1481 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1482 irq_context: 0 &mdev->graph_mutex &rq->__lock irq_context: 0 &mdev->graph_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1591 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1784 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1784 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#51 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 kn->active#49 &____s->seqcount#2 irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#52 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#52 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 kn->active#49 &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#53 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1665 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1665 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1666 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1666 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#26 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner irq_context: 0 (wq_completion)wg-kex-wg0#53 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1678 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1678 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1678 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1678 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1679 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1679 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1679 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1679 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2015 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#37 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work &p->pi_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1679 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1679 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1680 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1680 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1680 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1680 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1483 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1483 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1483 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1484 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#375 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#375 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#375 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#375 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq irq_context: 0 (wq_completion)nfc20_nci_cmd_wq irq_context: 0 (wq_completion)nfc20_nci_rx_wq irq_context: 0 (wq_completion)nfc20_nci_tx_wq irq_context: 0 (wq_completion)nfc24_nci_rx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#48 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#48 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc54_nci_cmd_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy34 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy33 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1681 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1681 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1682 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1682 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy33 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1683 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1682 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1683 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1683 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1683 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1684 irq_context: 0 sb_writers#5 pcpu_lock irq_context: 0 sb_writers#5 percpu_counters_lock irq_context: 0 sb_writers#5 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1683 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#31 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#16 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#31 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#32 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1687 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#16 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1691 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 key irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1695 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1695 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex minors_rwsem gdp_mutex kernfs_idr_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1696 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1696 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1696 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1695 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1697 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1697 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1697 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1697 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1697 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1696 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &sem->wait_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1698 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1698 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1698 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1698 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1698 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1697 irq_context: 0 &wq->mutex &obj_hash[i].lock irq_context: 0 &wq->mutex key irq_context: 0 &wq->mutex pcpu_lock irq_context: 0 &wq->mutex percpu_counters_lock irq_context: 0 &wq->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1486 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1486 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1486 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1486 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1486 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1487 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem &sem->wait_lock irq_context: 0 misc_mtx console_lock console_srcu console_owner_lock irq_context: 0 misc_mtx console_lock console_srcu console_owner irq_context: 0 misc_mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1698 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1699 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1699 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1699 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1700 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1700 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1700 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1700 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1700 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1700 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1700 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1699 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1701 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1701 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1701 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1701 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1701 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1701 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1700 irq_context: 0 lock map_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1487 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1487 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1487 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1488 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#377 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#377 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#377 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#377 irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 percpu_counters_lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#1488 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1488 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1703 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1703 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1703 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1702 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1489 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1489 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1489 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1489 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#378 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#378 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#378 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#378 irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq_wait->wait irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &__ctx->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 namespace_sem stock_lock irq_context: 0 &sbi->s_writepages_rwsem &__ctx->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock key irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1704 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1704 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1704 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1703 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1490 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1490 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1490 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1490 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1705 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1705 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1705 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1705 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1704 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#379 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#379 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#379 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#379 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1491 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1491 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1491 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1491 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1491 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1491 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1491 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1491 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1706 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1706 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1706 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1706 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1706 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1706 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1706 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock &vma->vm_lock->lock pcpu_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock percpu_counters_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1705 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1492 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1492 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1492 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1492 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1492 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1492 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1707 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1707 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1707 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1707 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1707 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1706 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &lruvec->lru_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1708 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1708 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1708 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1708 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1708 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1708 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1707 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1493 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1493 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1493 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#380 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#380 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#380 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#380 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#380 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1493 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1493 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1493 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#380 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1709 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1709 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1709 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1709 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1708 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock crngs.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1710 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1710 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1710 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1710 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1709 irq_context: 0 nfc_devlist_mutex dpm_list_mtx &rq->__lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1494 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1494 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1494 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1494 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1494 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1494 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1494 irq_context: 0 &type->i_mutex_dir_key#2 &xa->xa_lock#5 irq_context: 0 &type->i_mutex_dir_key#2 &xa->xa_lock#5 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#2 stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &meta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &obj_hash[i].lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 nf_sockopt_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1711 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1713 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1713 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1713 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1713 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1713 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1713 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1713 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1712 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1495 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1495 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1495 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1495 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1495 irq_context: 0 rlock-AF_KCM irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1714 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1714 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1714 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1714 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1713 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1496 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1590 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1496 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1496 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1496 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1496 irq_context: 0 &sb->s_type->i_mutex_key#10 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1496 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#381 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#381 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#381 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#381 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#381 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#381 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#381 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#381 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1715 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1715 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1715 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1715 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1715 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1714 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1714 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1715 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1717 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &lock->wait_lock irq_context: 0 kn->active#4 &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1723 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1723 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1722 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#3 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#3 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#53 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#53 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#53 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#53 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#53 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#269 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#269 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#269 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#269 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#269 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1724 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1724 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#6 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1502 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1502 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1502 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1723 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1502 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1502 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1502 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1502 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#387 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#387 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1502 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#42 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#42 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#42 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#42 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#42 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#387 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#387 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#42 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#42 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#387 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#54 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#54 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &rq->__lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#13 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &____s->seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#14 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#14 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#14 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#54 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#54 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#54 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#270 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#270 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#270 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#270 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#270 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#270 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#270 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1725 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1725 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1725 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1725 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1725 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1725 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1725 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1724 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#4 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#388 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#388 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#388 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#388 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1503 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1503 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1503 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1503 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1503 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1503 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#55 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#55 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#55 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#55 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#55 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#55 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#43 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#43 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#43 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#43 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#43 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#43 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1784 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1780 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy52 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1785 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1785 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1785 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1781 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#33 net_rwsem &list->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1786 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1786 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1786 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1782 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1787 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1787 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1787 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1787 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1772 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1773 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1787 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1783 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#46/1 &rq->__lock irq_context: 0 &type->s_umount_key#46/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#51 &n->list_lock irq_context: 0 kn->active#51 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#400 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#400 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#400 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#400 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#66 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#66 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#401 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#67 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#67 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#67 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#67 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#402 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#402 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1520 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1520 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#404 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#69 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#404 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#405 &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1521 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1521 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1521 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1738 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1738 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1739 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1523 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#408 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#70 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#70 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#70 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#70 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1524 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1524 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#389 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#389 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1504 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1504 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1504 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1504 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#271 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#271 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#389 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#271 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1726 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1742 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1743 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#410 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1743 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#410 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#410 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1526 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1526 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1526 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1526 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1745 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1745 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1745 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1527 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1527 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1745 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1746 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1746 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1746 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1746 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1746 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1746 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#411 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#411 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#411 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#411 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#411 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#411 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1528 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1747 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1747 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1747 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1747 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#271 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#271 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1726 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1747 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1746 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1529 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#271 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1529 &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1726 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1726 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1726 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1726 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1726 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1725 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1748 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1748 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1748 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1748 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1530 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1530 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1530 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1725 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1749 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1749 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1749 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1748 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_rx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#5 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#2 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_tx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#44 irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1531 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1531 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1531 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1531 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1531 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1750 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1750 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1750 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1750 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1750 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1749 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1749 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1749 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#44 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#44 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#8 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#54 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#54 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1532 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1532 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1532 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1751 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1751 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1751 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1751 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1751 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#286 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#286 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#286 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1750 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#286 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#286 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#286 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#412 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#412 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 irq_context: 0 (wq_completion)wg-crypt-wg0#29 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#412 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#412 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#412 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#412 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1752 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1751 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1533 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1533 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1533 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1533 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1533 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#21 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#71 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#71 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#71 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#71 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#71 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#71 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#71 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1753 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1753 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1753 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1753 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1752 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1534 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1534 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1534 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1534 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1534 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1534 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#287 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#287 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#287 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#287 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#287 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#287 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#287 irq_context: 0 kn->active#13 &n->list_lock irq_context: 0 kn->active#13 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#287 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#287 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#414 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#414 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#414 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#29 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1535 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1535 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1535 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1535 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1535 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#414 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#414 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#414 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1535 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#72 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#72 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1754 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1754 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1754 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#72 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_rx_wq#44 irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &base->lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#44 irq_context: 0 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#8 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#8 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#2 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#6 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#30 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#391 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#391 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#273 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#273 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#273 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#273 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#273 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#273 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#273 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#273 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#392 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#392 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#392 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#392 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#274 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#274 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#274 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#72 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#72 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1754 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1754 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1753 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1753 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1753 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1755 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1755 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1755 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1755 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1755 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1755 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1755 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1754 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1536 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1536 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1536 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1536 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1536 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1536 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1536 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1756 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1756 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1755 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1755 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1755 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#415 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#415 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#415 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1537 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1537 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1537 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1538 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1538 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1538 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1538 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1538 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1538 &obj_hash[i].lock irq_context: 0 pcpu_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1538 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1538 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) quarantine_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) remove_cache_srcu irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#29 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0#9 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci0#10 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#10 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &obj_hash[i].lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#416 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#416 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#416 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#416 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#416 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#416 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#416 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#416 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &dir->lock &obj_hash[i].lock irq_context: softirq rcu_callback &dir->lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1539 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1539 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1539 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1539 rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1539 irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#35 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex minors_rwsem &____s->seqcount#2 irq_context: 0 &dev->mutex minors_rwsem &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1759 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1759 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#500 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#500 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#24 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#27 &____s->seqcount#2 irq_context: 0 kn->active#27 &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &xa->xa_lock#4 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &k->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &base->lock irq_context: 0 sb_writers#3 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#15 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle bit_wait_table + i irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 irq_context: 0 (wq_completion)wg-crypt-wg2#15 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#30 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)wg-kex-wg2#30 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#55 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem stock_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1761 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1759 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1540 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1540 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1540 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1540 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1540 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1540 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#417 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#417 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#417 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#417 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1762 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1762 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1762 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1762 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1762 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1762 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1762 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1762 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock &wq irq_context: 0 nf_sockopt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#418 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#418 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1541 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1541 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1541 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#418 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#418 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#288 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#288 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#288 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#288 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#288 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#288 &rq->__lock irq_context: 0 pernet_ops_rwsem stock_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1764 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1764 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1764 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1764 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1764 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1761 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1761 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1761 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 strings &c->lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1542 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1542 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1542 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1542 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#419 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#419 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#419 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#419 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#419 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#419 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#419 irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1765 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1765 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1765 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1765 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1765 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1762 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1762 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1762 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1543 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1543 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1543 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1543 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#24 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#24 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 pernet_ops_rwsem krc.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &tn->node_list_lock irq_context: 0 pernet_ops_rwsem ebt_mutex irq_context: 0 pernet_ops_rwsem &xt[i].mutex irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1766 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1766 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1766 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1544 &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex.wait_lock irq_context: 0 pernet_ops_rwsem dev_base_lock irq_context: 0 pernet_ops_rwsem lweventlist_lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem napi_hash_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &fn->fou_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1546 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1546 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ipvs->sync_mutex irq_context: 0 pernet_ops_rwsem hwsim_radio_lock irq_context: 0 pernet_ops_rwsem &ent->pde_unload_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#422 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1547 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1547 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1547 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1547 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1770 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 pernet_ops_rwsem k-slock-AF_INET#2 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem &sn->gssp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#25 irq_context: 0 pernet_ops_rwsem &sn->gssp_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#25 irq_context: 0 pernet_ops_rwsem cache_list_lock &cd->hash_lock irq_context: 0 pernet_ops_rwsem (&net->can.stattimer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#26 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#26 irq_context: 0 pernet_ops_rwsem xfrm_state_gc_work irq_context: 0 pernet_ops_rwsem &net->xfrm.xfrm_state_lock irq_context: 0 pernet_ops_rwsem &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#27 irq_context: 0 pernet_ops_rwsem &ht->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#27 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 slock-AF_INET#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 rlock-AF_INET irq_context: 0 pernet_ops_rwsem key#24 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &(&priv->bus_notifier)->rwsem kfence_freelist_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1771 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1772 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1772 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1772 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1772 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1772 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1769 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1769 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1773 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1773 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1773 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1548 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1548 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1548 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_ct_proto_mutex.wait_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1548 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->mutex &dev->mutex register_mutex#6 quarantine_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: softirq (&app->join_timer) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&app->join_timer) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&app->join_timer) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock pool_lock irq_context: 0 namespace_sem &n->list_lock irq_context: 0 namespace_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#15 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1290 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1290 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#29 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1432 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1432 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1433 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1433 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1433 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1433 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1290 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1290 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1290 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1290 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1290 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1290 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1290 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1291 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock key irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock pcpu_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock percpu_counters_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1434 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1434 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1434 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1291 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1291 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1291 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1291 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1291 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1291 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1291 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#356 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#356 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#356 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1292 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#356 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1292 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#356 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1292 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#356 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#356 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#356 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#356 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#356 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_node_0 irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)bond0#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1282 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#25 irq_context: 0 &type->s_umount_key#47 fill_pool_map-wait-type-override &c->lock irq_context: 0 &type->s_umount_key#47 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &type->s_umount_key#47 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#47 fill_pool_map-wait-type-override pool_lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#50 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#49 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &c->lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_SIT irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#49 remove_cache_srcu irq_context: 0 kn->active#49 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#49 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci0#7 irq_context: 0 (wq_completion)hci0#7 &rq->__lock irq_context: 0 (wq_completion)hci0#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 kn->active#49 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 kn->active#49 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#49 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#49 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex.wait_lock irq_context: 0 ebt_mutex &____s->seqcount#2 irq_context: 0 ebt_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#25 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#50 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#49 irq_context: 0 (wq_completion)wg-kex-wg0#59 irq_context: 0 pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#30 &rq->__lock irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 &hdev->req_lock (wq_completion)hci5#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#49 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: softirq rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 key#24 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 vmap_area_lock irq_context: 0 &u->iolock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#59 irq_context: 0 &u->iolock &rcu_state.expedited_wq irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 &dentry->d_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 sb_writers#12 mount_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 sb_writers#12 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 sb_writers#12 &sb->s_type->i_lock_key#27 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 sb_writers#12 &wb->list_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#17 sb_writers#12 &wb->list_lock &sb->s_type->i_lock_key#27 irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem net_rwsem net_rwsem.wait_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1685 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1685 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1685 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1685 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1684 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#32 irq_context: 0 sb_writers#3 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1686 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1686 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1686 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1686 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1686 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1685 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem net_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex quarantine_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: softirq rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex nf_hook_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_hook_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#60 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#60 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#61 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#62 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#62 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1282 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1282 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#62 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#63 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#63 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#63 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &n->list_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#64 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#64 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#64 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#65 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#65 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#65 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#66 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#66 irq_context: 0 &p->lock &of->mutex kn->active#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#67 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#67 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#68 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#68 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#68 irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex gdp_mutex.wait_lock irq_context: 0 rtnl_mutex gdp_mutex.wait_lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#32 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#68 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#69 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#69 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#69 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#69 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#69 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#69 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#70 irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond0#16 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#32 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1806 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1806 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1806 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1806 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1806 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1802 irq_context: 0 ebt_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 ebt_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1807 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1807 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1807 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1807 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1807 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1803 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1803 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex ovs_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1560 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1560 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1560 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1559 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1559 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#428 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#428 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#428 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1808 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1808 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1808 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#428 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#428 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1808 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#428 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1808 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1804 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#293 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#293 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#293 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#39 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#36 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#15 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#293 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#293 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#293 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#293 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#293 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1809 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1809 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1809 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1590 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1590 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1590 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#306 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 put_task_map-wait-type-override#3 &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override#3 &pcp->lock &zone->lock irq_context: 0 put_task_map-wait-type-override#3 &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1434 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1434 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1435 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1435 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1435 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1435 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &meta->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1435 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1435 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1436 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#59 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#52 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#54 irq_context: 0 put_task_map-wait-type-override#3 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1809 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1809 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1805 irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#55 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1561 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1561 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1561 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1561 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1560 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1560 irq_context: 0 binderfs_minors_mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock &n->list_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &port_dev->status_lock hcd->address0_mutex &queue->lock &n->list_lock &c->lock irq_context: 0 binderfs_minors_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#429 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#429 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#429 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1562 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1562 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1561 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1561 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1811 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#430 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1811 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1811 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1807 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#430 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1813 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 pcpu_alloc_mutex key irq_context: 0 pcpu_alloc_mutex percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 rcu_node_0 irq_context: 0 rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 kn->active#47 &n->list_lock irq_context: 0 kn->active#47 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1425 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1425 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1425 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1425 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1282 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1282 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1282 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1283 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1283 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1283 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1283 irq_context: 0 &wq->mutex &rcu_state.expedited_wq irq_context: 0 &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1283 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1284 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1436 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1292 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1292 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1292 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1292 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1292 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1293 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#15 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1437 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1437 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1426 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1426 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1426 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1427 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1427 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1427 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1427 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1293 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1293 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1284 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1284 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1284 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1285 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1285 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1285 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1427 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1427 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1427 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &rcu_state.expedited_wq irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1427 irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1428 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1428 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1285 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1285 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1285 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1285 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1285 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1286 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1293 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1294 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1437 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1438 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1438 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1438 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1294 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1294 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1294 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1294 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1294 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1295 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#357 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#357 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#357 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#357 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#357 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#357 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1295 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1295 &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1295 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1295 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1296 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1438 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1438 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1439 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1439 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#255 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#255 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#255 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#255 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1429 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1429 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1429 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#355 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#355 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#355 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#355 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#355 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1286 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1286 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1286 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1286 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1287 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#254 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#254 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#254 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1429 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1429 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1430 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1430 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1430 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1430 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1287 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1287 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1287 irq_context: 0 (wq_completion)wg-kex-wg0#43 irq_context: 0 (wq_completion)wg-kex-wg0#43 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1813 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1814 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1814 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1814 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1810 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1815 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1815 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1815 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#29 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1816 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1816 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1816 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1813 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1567 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1567 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1816 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1816 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1812 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1817 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1817 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1817 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1817 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1817 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1817 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1818 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1818 &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->pipefs_sb_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1819 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1819 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1819 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1819 irq_context: 0 (wq_completion)netns net_cleanup_work &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1815 irq_context: 0 put_task_map-wait-type-override#3 stock_lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#306 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#305 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#305 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#305 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#59 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#59 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#305 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1846 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1846 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1846 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1846 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1569 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1569 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1569 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#113 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#70 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#70 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1842 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#71 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#71 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#71 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#72 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#72 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#72 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1842 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#73 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1842 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1592 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1592 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#73 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#73 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#73 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#74 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#74 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#74 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#75 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#75 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#76 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#76 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#76 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#76 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#77 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#446 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#446 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#69 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#70 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#70 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#71 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#71 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#72 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1592 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1591 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1591 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#89 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#99 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#446 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#446 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1591 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#60 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#60 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#60 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#60 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#86 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1821 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1821 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1821 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1817 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1817 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#92 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#96 irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#73 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#60 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1822 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1822 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1822 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#98 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#99 irq_context: 0 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1824 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1824 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1823 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#100 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1823 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1823 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1823 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1823 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1824 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#74 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#76 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#77 irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1824 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1573 irq_context: 0 sched_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1573 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#103 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1826 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#26 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#26 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 irq_context: 0 (wq_completion)wg-kex-wg1#53 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#53 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#54 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#53 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#53 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#54 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1826 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#104 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1574 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#105 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#106 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1829 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1829 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1576 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1830 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1830 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1577 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1831 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1831 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1831 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1577 &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1578 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1577 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1579 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1579 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#435 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1580 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1580 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1835 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#307 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#307 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#307 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#307 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#306 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#84 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#84 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#84 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#84 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#84 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#84 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#61 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#61 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#61 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#61 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#308 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#308 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#308 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1847 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#308 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#307 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1847 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#307 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1843 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#85 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#85 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#85 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#85 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#85 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#447 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#447 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#447 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#447 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#447 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#447 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#447 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#447 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#116 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1848 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1848 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1848 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1844 irq_context: 0 &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#448 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#448 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#448 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#448 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#448 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#24 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1592 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1592 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1592 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#449 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#449 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#449 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#449 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#449 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#449 irq_context: 0 (work_completion)(&local->timeout_work) &rq->__lock irq_context: 0 (work_completion)(&local->timeout_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1594 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1594 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1594 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#88 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#455 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#125 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1854 irq_context: 0 (wq_completion)wg-crypt-wg0#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#313 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#313 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#313 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#313 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#312 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#312 irq_context: 0 &type->i_mutex_dir_key#4 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#65 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#65 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1855 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#127 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#456 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1603 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1603 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1603 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1603 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1602 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1602 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#128 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1603 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1603 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1603 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1603 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1603 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1856 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1856 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#130 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1856 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1852 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1852 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1852 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#131 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#314 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#314 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#313 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#313 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#313 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#313 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#457 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#457 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#457 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#457 irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#457 irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#132 irq_context: 0 &p->lock percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#315 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#315 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#314 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#134 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1605 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1605 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1604 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1858 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1858 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1858 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#135 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1858 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#459 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &base->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1606 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#90 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#90 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#90 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#315 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#137 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1859 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1859 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1859 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1859 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1855 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1607 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1607 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1607 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1607 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1606 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1860 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1860 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1860 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1860 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1860 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1860 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1860 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1856 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1608 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#461 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#461 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#461 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#461 &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock key irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex rcu_read_lock &rq->__lock irq_context: 0 wq_pool_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1861 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1609 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#317 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#316 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#316 irq_context: 0 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1610 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1610 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1609 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1611 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1611 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1611 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1610 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#318 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1610 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#318 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#318 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#318 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#317 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#317 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#317 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#317 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1610 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1610 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#66 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#66 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#66 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#462 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#462 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#462 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#462 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#462 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#462 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1863 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1863 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1863 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1863 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1859 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1612 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1864 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1864 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1864 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1864 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1864 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1861 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1866 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1866 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1866 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1866 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#464 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1867 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1867 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1867 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1867 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1868 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1868 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1868 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1864 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1864 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1864 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#320 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1869 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1869 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1869 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1869 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1869 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1871 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1617 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1617 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1616 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1874 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1874 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1875 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1875 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1875 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1875 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1875 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1875 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#29 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1876 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1876 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1876 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1876 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1876 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1872 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex key irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)bond0#31 irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1877 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1877 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1877 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1877 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1877 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1873 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1873 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1878 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1878 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1878 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1878 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#29 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1879 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1879 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1879 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1879 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#61 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1881 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1885 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1885 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#139 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#139 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#140 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#140 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1886 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1886 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1886 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1886 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1886 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1882 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy55 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1887 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1887 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1887 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1887 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1887 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1888 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1888 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1888 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1888 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1888 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1888 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1884 irq_context: 0 put_task_map-wait-type-override &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override pool_lock#2 irq_context: 0 put_task_map-wait-type-override stock_lock irq_context: 0 put_task_map-wait-type-override per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#26 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1889 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1889 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1889 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1885 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1885 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1885 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1890 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1890 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1890 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1891 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1618 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1887 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1618 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1887 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1887 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#467 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#467 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#93 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#93 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#93 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#467 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#93 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#93 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#93 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#467 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#467 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#467 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#321 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#321 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#320 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#320 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1893 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1893 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1892 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1892 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1888 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1888 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1888 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1620 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1620 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1620 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1620 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1619 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1619 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1619 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1619 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1894 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1894 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1893 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1889 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1889 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1889 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1895 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1895 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1891 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1897 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1897 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1897 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1898 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1898 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1898 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1898 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1896 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1896 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1896 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1892 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1621 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1621 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1620 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1620 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1620 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1620 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1897 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1897 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1897 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1893 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1893 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1893 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1622 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1622 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1622 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1621 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1621 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1900 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1900 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1898 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1894 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1623 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1623 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1622 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1622 irq_context: 0 &hdev->req_lock (wq_completion)hci4#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1901 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1901 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1899 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1895 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1624 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1624 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1624 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1624 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1623 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1623 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1623 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1623 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex deferred_probe_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#12 irq_context: 0 (wq_completion)hci4#12 &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &base->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#11 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1902 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1903 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &f->f_pos_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1902 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1902 &rq->__lock irq_context: 0 link_idr_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1902 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#498 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)hci2#15 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem wq_pool_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#34 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#498 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#23 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#23 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#6 irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#358 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1996 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1996 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_rx_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1955 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1955 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1955 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1955 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1951 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1946 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1670 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1670 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#501 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#501 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#501 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#501 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#499 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#499 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx rcu_node_0 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rcu_state.expedited_wq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#18 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#15 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#18 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1669 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1669 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#499 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)nfc4_nci_tx_wq#499 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#339 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#339 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#338 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#338 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#502 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#502 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#500 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#500 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#500 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#79 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#79 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)nfc4_nci_tx_wq#500 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#340 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#340 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#339 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#339 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1378 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1237 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1237 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#74 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1956 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1956 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1956 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1956 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#74 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1237 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1238 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1378 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#249 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1378 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1378 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#74 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1378 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#249 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#249 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#249 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1379 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1379 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1379 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1379 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#249 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#249 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#249 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1238 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1238 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#38 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1380 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1380 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1380 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#19 irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#74 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1239 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1239 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1239 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1240 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1380 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0#11 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &meta->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1380 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM slock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM nr_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM rlock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_NETROM irq_context: 0 uts_sem &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 uts_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1957 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1957 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1957 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#340 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1672 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1672 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1672 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1672 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#503 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#503 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#504 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci0#12 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1959 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1960 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1956 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1956 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1992 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1987 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1710 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1710 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1710 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1710 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1709 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1709 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1709 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1709 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#523 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &meta->lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#523 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#525 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#525 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#12 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#525 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#35 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#35 &devlink_port->type_lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy60 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy60 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#525 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#524 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#526 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#20 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1711 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1711 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1710 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1710 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1997 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1997 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1993 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1988 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1998 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1998 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1994 irq_context: 0 (wq_completion)events_power_efficient &rq->__lock irq_context: 0 (wq_completion)events_power_efficient &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1989 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1712 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1712 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1711 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1711 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1711 &rq->__lock irq_context: 0 loop_validate_mutex &cfs_rq->removed.lock irq_context: 0 loop_validate_mutex &obj_hash[i].lock irq_context: 0 loop_validate_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1711 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy59 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy59 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1999 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1995 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1995 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1995 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1990 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1713 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1713 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1712 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1712 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#527 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#527 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#525 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#525 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy38 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex dpm_list_mtx irq_context: 0 (wq_completion)wg-crypt-wg0#33 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy37 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#522 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#361 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#361 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#361 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#361 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#359 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#359 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#116 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#116 &rq->__lock irq_context: 0 (wq_completion)bond0#20 &rq->__lock irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1381 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1381 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1381 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1381 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#429 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#429 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1810 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1810 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1952 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1952 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1952 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1952 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1947 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1947 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1947 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1671 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1671 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1671 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1953 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1953 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1953 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1671 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1670 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1670 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1948 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#341 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#341 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#341 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#341 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1810 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1810 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1810 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1806 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2010 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2010 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2010 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1997 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1997 &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1806 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2010 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2010 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2006 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2006 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2006 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2001 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2007 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2003 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &p->pi_lock irq_context: 0 kn->active#13 remove_cache_srcu irq_context: 0 kn->active#13 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#13 remove_cache_srcu &c->lock irq_context: 0 kn->active#13 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#13 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1723 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2003 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1998 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1998 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1998 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1722 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1722 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1723 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2003 irq_context: 0 kn->active#12 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy64 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1722 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1722 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#536 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#536 rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1474 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1474 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1474 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1475 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1628 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1629 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1629 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1475 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1475 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1475 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1476 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#46 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#45 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#45 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#46 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#41 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &n->list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#42 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &base->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#67 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#67 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#67 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#68 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2003 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1999 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1994 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1718 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1718 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1716 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1715 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#531 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#531 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#531 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#529 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#526 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1723 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#534 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#534 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#534 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#532 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#529 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#529 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#537 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#537 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1720 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1720 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1720 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1720 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1718 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1717 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#365 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#365 irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#535 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#535 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#535 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1721 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1721 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1721 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#535 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#533 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#530 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#530 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1806 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#116 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#116 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#115 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#115 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#532 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#68 irq_context: 0 (wq_completion)wg-kex-wg0#68 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#68 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#38 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#38 &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#38 &devlink_port->type_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1723 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1721 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1721 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1721 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1720 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1720 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1720 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2011 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2011 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2007 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2002 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy76 irq_context: 0 misc_mtx &rq->__lock &obj_hash[i].lock irq_context: 0 misc_mtx &rq->__lock &base->lock irq_context: 0 misc_mtx &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1724 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1724 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1724 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1724 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1722 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1722 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1721 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1721 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1721 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2012 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2012 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2012 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2012 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2012 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2012 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2008 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2008 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#79 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#79 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2008 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2003 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#79 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#79 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#538 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1725 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1725 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#536 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#533 irq_context: 0 nfc_devlist_mutex gdp_mutex gdp_mutex.wait_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1723 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1722 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#367 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#367 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#365 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2001 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2001 irq_context: 0 (wq_completion)wg-kex-wg1#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#35 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#18 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 lweventlist_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1997 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1992 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#31 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#31 irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#365 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy76 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#118 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#118 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#117 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2002 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#147 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2013 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2013 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2009 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2009 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2009 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2004 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1726 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1726 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1721 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1726 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1724 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1723 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2014 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2014 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2014 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2010 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2010 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2010 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2005 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy75 irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy75 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2015 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2015 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2015 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2015 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2011 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2006 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1727 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1727 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1727 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1727 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1725 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1724 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#539 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#539 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#539 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#539 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#537 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#534 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#534 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#39 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg1#39 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#40 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#31 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#73 irq_context: 0 cb_lock rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#71 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1719 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#147 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#147 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#74 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2010 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#371 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2010 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2010 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#371 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#371 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#543 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2002 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2002 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1998 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1993 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1717 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1717 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1717 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1717 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1715 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1714 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1714 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#34 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#363 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#363 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#363 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#543 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#541 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#541 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#361 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#361 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#361 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#361 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#530 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#538 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#41 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#530 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#41 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#530 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#41 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#38 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2016 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#530 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#528 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#525 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#525 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#80 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#80 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#80 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2005 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2005 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#363 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#363 rcu_node_0 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#363 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#363 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1719 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1718 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#366 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#366 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#364 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1720 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1719 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1719 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#80 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#80 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#80 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#80 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2016 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2012 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#80 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#536 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#536 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2007 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2007 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 nl_table_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#340 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#340 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#340 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#430 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#430 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#105 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#75 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#105 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#75 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#105 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#105 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#430 &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#430 irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1812 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1812 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2007 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1728 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1728 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1812 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1726 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1725 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#540 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#540 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#538 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1808 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#538 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#535 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#41 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#41 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#39 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#34 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#68 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#36 irq_context: 0 (wq_completion)wg-crypt-wg2#34 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#35 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#75 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#75 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#105 irq_context: 0 (wq_completion)wg-crypt-wg1#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#75 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#75 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#37 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#37 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#2020 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#83 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#83 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#83 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#37 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#83 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1731 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1731 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2021 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2021 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1731 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1729 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1728 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1728 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1728 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2016 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2011 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2011 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2011 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#544 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#544 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1809 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#431 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#431 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1563 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1563 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1562 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#431 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#431 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1562 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#294 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#294 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#294 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#294 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#294 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc12_nci_rx_wq#6 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc12_nci_cmd_wq#6 irq_context: 0 misc_mtx (wq_completion)nfc12_nci_cmd_wq#7 irq_context: 0 misc_mtx (wq_completion)nfc12_nci_cmd_wq#7 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc12_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#25 irq_context: 0 prog_idr_lock &obj_hash[i].lock irq_context: 0 prog_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#24 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#24 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#82 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#82 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#82 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#82 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#82 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#82 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#82 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#82 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#40 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#40 &rq->__lock irq_context: 0 lock prog_idr_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#40 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 net_rwsem irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1240 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1240 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1240 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1240 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1241 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1381 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1381 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1382 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1382 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1382 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#294 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#294 &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1814 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1814 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1814 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1814 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1564 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1564 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1964 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#40 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1964 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1964 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#40 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1964 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#38 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#38 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#39 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#40 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1964 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1960 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1965 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1965 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#368 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#368 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#366 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#366 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#20 irq_context: 0 (wq_completion)wg-crypt-wg0#20 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#40 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &cookie->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#544 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#542 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#366 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#366 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#40 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#120 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#120 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#120 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 cb_lock genl_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#119 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#541 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#541 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#541 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#539 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#539 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#539 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2022 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2022 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2022 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex &n->list_lock irq_context: 0 ebt_mutex &n->list_lock &c->lock irq_context: 0 ebt_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#75 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 sb_writers#3 &sem->wait_lock irq_context: 0 sb_writers#3 &p->pi_lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#20 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#75 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#536 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#37 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &fs->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &fs->lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &dentry->d_lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#38 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 nfc_devlist_mutex remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#76 irq_context: 0 (wq_completion)wg-kex-wg0#76 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2022 irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2017 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#38 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2017 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2017 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &br->hash_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#75 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2012 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1732 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1732 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1732 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 rcu_read_lock tasklist_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock tasklist_lock &sighand->siglock irq_context: 0 rcu_read_lock tasklist_lock &sighand->siglock &p->pi_lock irq_context: 0 rcu_read_lock tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#545 irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#40 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#39 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2019 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_node_0 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg1#41 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock stock_lock irq_context: 0 rtnl_mutex &wg->device_update_lock key irq_context: 0 rtnl_mutex &wg->device_update_lock pcpu_lock irq_context: 0 rtnl_mutex &wg->device_update_lock percpu_counters_lock irq_context: 0 rtnl_mutex &wg->device_update_lock pcpu_lock stock_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#19 irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)wg-kex-wg2#37 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#37 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#37 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#38 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#59 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#25 irq_context: 0 (wq_completion)wg-kex-wg2#60 irq_context: 0 (wq_completion)wg-kex-wg0#60 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#59 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#60 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#50 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#30 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1714 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1714 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#1713 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1715 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#1715 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)bond0#38 irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#38 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#38 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#38 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#38 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#38 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#38 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#38 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 cbs_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#38 &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#38 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 lock irq_context: 0 rtnl_mutex team->team_lock_key#38 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &root->kernfs_rwsem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#38 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#38 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#38 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#38 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#38 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#38 console_lock console_srcu console_owner &port_lock_key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#38 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#117 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#117 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#117 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#117 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#116 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#116 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#38 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#38 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#38 &____s->seqcount#2 irq_context: 0 rtnl_mutex team->team_lock_key#38 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#38 quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2000 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2000 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#24 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#47 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2000 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2000 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1996 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1991 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex rcu_node_0 irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &base->lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1716 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1716 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1716 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#38 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1716 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1714 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1713 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#528 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#528 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#526 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#523 irq_context: 0 (wq_completion)wg-crypt-wg0#19 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1732 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2009 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1730 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1730 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#37 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#362 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#362 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#19 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#76 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#362 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#15 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci2#8 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#42 irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#41 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#21 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#42 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#42 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#42 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#41 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#21 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#19 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 quarantine_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1404 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1405 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_internal &n->list_lock irq_context: 0 sb_internal &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy40 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1367 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1226 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1226 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1226 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1226 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1226 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy39 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1227 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1227 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1227 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1227 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1227 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1227 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1228 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1369 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1228 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1228 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1229 irq_context: 0 sb_writers#4 oom_adj_mutex.wait_lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1369 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1369 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1370 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1370 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1229 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1229 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1229 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1230 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1382 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1241 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1241 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1241 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1241 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1241 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1242 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1242 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1242 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1243 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1243 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1243 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1370 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1371 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1371 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1230 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1230 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1230 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1243 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1243 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1231 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1243 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1385 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1244 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1244 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1244 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1372 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1372 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1231 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1231 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1231 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1232 irq_context: 0 &uhid->devlock irq_context: 0 &uhid->devlock &rq->__lock irq_context: 0 &uhid->devlock &mm->mmap_lock irq_context: 0 &uhid->devlock fs_reclaim irq_context: 0 &uhid->devlock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &uhid->devlock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &uhid->devlock &pcp->lock &zone->lock irq_context: 0 &uhid->devlock &____s->seqcount irq_context: 0 &uhid->devlock &x->wait#9 irq_context: 0 &uhid->devlock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1553 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1554 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1405 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1405 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1406 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1406 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy43 irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#9 irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) rcu_node_0 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#9 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#10 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 &uhid->devlock rcu_read_lock &pool->lock irq_context: 0 &uhid->devlock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &uhid->devlock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &uhid->devlock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) dquirks_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) free_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) init_mm.page_table_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) pin_fs_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) bus_type_sem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &uhid->report_wait irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &hdrv->dyn_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex device_links_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex pinctrl_list_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex pinctrl_maps_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex semaphore->lock#3 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rq->__lock irq_context: 0 (work_completion)(&uhid->worker) irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex dquirks_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex free_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex init_mm.page_table_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) subsys mutex#77 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#77 irq_context: 0 &hdev->debug_wait irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1372 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1372 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1373 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1373 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1373 irq_context: 0 vmap_purge_lock &cfs_rq->removed.lock irq_context: 0 vmap_purge_lock &obj_hash[i].lock irq_context: 0 vmap_purge_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1232 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1232 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &uhid->devlock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1232 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1232 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1232 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1233 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) console_owner irq_context: 0 (wq_completion)nfc4_nci_rx_wq#346 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#346 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex console_owner irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#247 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#247 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#247 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#247 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#247 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1373 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1373 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1373 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) purge_vmap_area_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1374 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1374 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1233 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1233 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1233 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1233 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1233 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1233 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1233 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1234 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1245 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &x->wait#9 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem bus_type_sem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem req_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &x->wait#11 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1374 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1374 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1374 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1375 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1375 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1234 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1234 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1234 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1234 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1234 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1235 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1376 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1376 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1376 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1376 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1376 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1235 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1235 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1235 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1236 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#347 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_node_0 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#347 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#347 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#248 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#248 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#248 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1377 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1236 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1236 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1236 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1236 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1236 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1237 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &k->k_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1385 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem subsys mutex#78 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1385 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem subsys mutex#78 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem subsys mutex#78 &k->k_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1386 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1386 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1386 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1386 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex semaphore->lock#3 irq_context: 0 &dev->mutex minors_rwsem irq_context: 0 &dev->mutex minors_rwsem &k->list_lock irq_context: 0 &dev->mutex minors_rwsem &k->k_lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex minors_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex minors_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex minors_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex minors_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 &dev->mutex minors_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex minors_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem pool_lock#2 irq_context: 0 &dev->mutex minors_rwsem &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex minors_rwsem req_lock irq_context: 0 &dev->mutex minors_rwsem &p->pi_lock irq_context: 0 &dev->mutex minors_rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex minors_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem &x->wait#11 irq_context: 0 &dev->mutex minors_rwsem &rq->__lock irq_context: 0 &dev->mutex minors_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&barr->work) pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1245 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1245 irq_context: 0 &dev->mutex minors_rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex minors_rwsem subsys mutex#78 irq_context: 0 &dev->mutex minors_rwsem subsys mutex#78 &k->k_lock irq_context: 0 &dev->mutex minors_rwsem subsys mutex#78 &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex minors_rwsem &x->wait#9 irq_context: 0 &dev->mutex minors_rwsem dpm_list_mtx irq_context: 0 &dev->mutex minors_rwsem &dev->power.lock irq_context: 0 &dev->mutex minors_rwsem deferred_probe_mutex irq_context: 0 &dev->mutex minors_rwsem device_links_lock irq_context: 0 &dev->mutex minors_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex minors_rwsem &c->lock irq_context: 0 &dev->mutex minors_rwsem &n->list_lock irq_context: 0 &dev->mutex minors_rwsem &n->list_lock &c->lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex minors_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem gdp_mutex irq_context: 0 &dev->mutex minors_rwsem gdp_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex minors_rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex minors_rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex minors_rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex minors_rwsem gdp_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex &uhid->qlock irq_context: 0 &dev->mutex &uhid->qlock &uhid->waitq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1245 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1246 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1246 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1246 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1246 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1246 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1246 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1247 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1387 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1387 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1388 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1388 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1247 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1247 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1247 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1248 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#349 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#349 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#349 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#349 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#349 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#250 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#250 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#250 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#250 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#250 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1389 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1248 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1248 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1248 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1248 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1248 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1249 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1390 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1390 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#350 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#350 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1249 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#350 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1249 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1249 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#350 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#350 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#251 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#251 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1249 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1249 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#251 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1250 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1250 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1250 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#251 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1390 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1391 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1391 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1250 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1250 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1250 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1251 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1392 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1251 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1251 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1251 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1251 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1251 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1251 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1251 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1252 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1393 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1393 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1393 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1252 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1252 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1252 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1252 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1252 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1253 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#351 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#351 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1393 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1393 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1393 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1394 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1394 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1394 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1395 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1395 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1253 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1254 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1254 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1396 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1254 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1255 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#352 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#352 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#352 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#252 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#252 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#252 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#352 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#352 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#252 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#252 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#252 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1255 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1255 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1255 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1255 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1255 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1255 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1255 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1256 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1256 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1256 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1398 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1398 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1256 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1256 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1256 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1256 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1256 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1257 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1257 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1257 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1257 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1257 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1257 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1257 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1257 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1258 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1258 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1258 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1259 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1259 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1259 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1259 irq_context: 0 cgroup_threadgroup_rwsem stock_lock irq_context: 0 cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1259 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1260 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1260 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1404 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1404 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1262 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1262 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1262 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1262 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1262 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1262 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1262 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1263 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1404 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1405 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1405 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1263 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1263 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1263 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1263 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1263 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1264 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1264 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1264 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1265 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1265 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#353 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#353 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#353 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#353 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#353 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1406 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#253 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#253 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#253 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#253 irq_context: 0 (wq_completion)events free_ipc_work quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1406 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1406 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1407 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1407 irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1265 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1265 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1265 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1266 irq_context: 0 (wq_completion)hci4#5 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &n->list_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock quarantine_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1408 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1408 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1266 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1266 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1266 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1266 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1266 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1266 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1266 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1267 irq_context: 0 misc_mtx nl_table_wait.lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#354 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#354 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#354 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#354 irq_context: 0 &wq->mutex rcu_node_0 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#354 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1408 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1408 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1409 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1409 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1409 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1409 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1267 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1267 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1267 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1268 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#43 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#43 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1410 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1410 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 irq_context: 0 (wq_completion)wg-kex-wg0#44 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1411 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1411 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1411 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#43 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1268 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1268 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1268 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1268 irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1268 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#44 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#44 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1269 irq_context: 0 (wq_completion)wg-kex-wg2#44 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-crypt-wg1#22 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#22 irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#44 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem stock_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem key irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1411 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1411 irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1412 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1412 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1269 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1269 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1269 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1269 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1269 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1270 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1270 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1270 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#43 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#22 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1413 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1413 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1270 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1270 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1270 irq_context: 0 (wq_completion)wg-crypt-wg0#22 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#21/1 stock_lock irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 &c->lock irq_context: 0 &type->s_umount_key#21/1 &xa->xa_lock#5 &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1271 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1439 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1439 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1439 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1440 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1440 &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1296 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1296 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1296 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1296 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1296 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1297 irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &n->list_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->filter_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1440 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1441 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1441 irq_context: 0 &type->s_umount_key#21/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1441 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1298 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1298 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1298 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1299 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1299 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1299 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1300 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1444 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1444 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1444 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1300 irq_context: 0 (wq_completion)bond0#22 &rq->__lock irq_context: 0 (wq_completion)bond0#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1413 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1413 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1414 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1414 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1414 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1414 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1271 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1271 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1271 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1271 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1271 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1272 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1300 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1300 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1300 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1300 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1301 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1301 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1301 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1302 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1302 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1303 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1303 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1303 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1303 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1304 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1304 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1304 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1304 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1304 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1305 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1449 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1449 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1449 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1449 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1305 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1305 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1305 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1306 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1307 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1307 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1307 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1450 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1451 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1451 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#256 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#256 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#358 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#358 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#358 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#358 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#358 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#256 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#358 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1308 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1308 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1308 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1308 irq_context: 0 (wq_completion)events free_ipc_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1309 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#11 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1453 irq_context: 0 kn->active#11 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1453 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1453 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1309 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1309 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1309 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1309 &rq->__lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1407 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1407 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1407 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1407 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1408 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1556 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1556 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1556 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1557 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1557 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1408 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1408 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1408 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1408 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1409 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1558 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1558 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1558 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1409 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1409 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1409 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1410 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1410 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1410 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1410 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1410 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1411 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1559 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1560 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#17 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1560 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1560 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1560 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1411 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1411 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1411 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1411 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1412 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1412 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex rcu_read_lock stock_lock irq_context: 0 &xt[i].mutex rcu_read_lock key irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1414 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1414 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1414 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1415 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1415 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1415 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1415 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1272 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1272 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1272 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&local->sdreq_timeout_work) pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1273 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1273 irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#22 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1273 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1273 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1273 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1274 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1274 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1274 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1275 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1275 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1275 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#10 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1309 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1309 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1309 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1309 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1310 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1310 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1311 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1311 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1311 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1311 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1312 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1456 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1312 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1312 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1312 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1312 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1313 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1313 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1313 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1313 irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex.wait_lock irq_context: 0 &type->s_umount_key#21/1 &p->pi_lock irq_context: 0 &type->s_umount_key#21/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#21/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#21/1 &rq->__lock irq_context: 0 &type->s_umount_key#21/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1314 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1314 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1314 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1315 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1459 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1459 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1459 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1315 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1315 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1315 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1315 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1315 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1315 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1315 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1315 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1316 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1460 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1460 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1316 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1316 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1316 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1316 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1317 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1317 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1460 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1317 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1317 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1317 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1317 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1318 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1318 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1318 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1318 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1318 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1318 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1319 irq_context: 0 ebt_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 jbd2_handle stock_lock irq_context: 0 sb_writers#3 jbd2_handle pcpu_lock stock_lock irq_context: 0 wq_pool_attach_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1462 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1462 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1463 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1319 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1319 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1319 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1319 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1320 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1320 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1320 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1464 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1320 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1320 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1320 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1320 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1320 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1320 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1320 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1321 irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#16 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1465 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1465 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1321 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1321 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1321 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1321 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1321 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1322 irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1322 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1322 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1323 irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#18 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1323 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1323 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1466 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1467 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1467 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1323 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1323 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1323 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1324 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1324 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1324 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1468 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1468 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1324 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1324 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1324 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1324 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1325 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1325 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1325 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1325 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1325 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1325 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1325 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1326 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1326 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1326 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &base->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1326 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1326 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1326 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1327 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1470 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1471 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1471 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1327 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1327 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1327 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1327 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1327 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1327 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1328 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1471 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1472 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1472 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1328 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1328 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1328 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1328 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1328 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1329 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1329 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1329 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1473 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1473 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1329 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1329 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1329 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1330 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1330 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1330 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1473 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 key irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1418 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1418 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1418 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1418 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1418 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1419 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1419 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1419 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1419 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1419 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1419 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1276 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1276 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1276 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1276 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1276 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1277 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1474 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1419 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1420 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1420 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1420 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1474 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1330 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1330 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1331 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1331 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1331 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1474 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1474 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1474 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1475 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1475 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1475 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1475 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1475 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1475 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1331 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1331 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1331 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1420 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1277 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1277 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1277 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1277 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1278 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#359 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#359 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#359 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1332 irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#359 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#359 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#359 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#359 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#257 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#257 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#257 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#257 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1332 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1333 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#48 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#48 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#47 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy45 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1420 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1420 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock percpu_counters_lock irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 &xt[i].mutex rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#369 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#369 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#369 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#369 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1560 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1560 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1560 irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock quarantine_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1561 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1561 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1561 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1562 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1562 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1412 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1412 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1412 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1413 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1413 &rq->__lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &dentry->d_lock &wq irq_context: 0 tomoyo_ss rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1413 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1413 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1414 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1564 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1565 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1565 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1415 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1415 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1421 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 tomoyo_ss &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#24 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1435 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1435 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1435 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1436 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1587 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1436 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1436 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1436 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1588 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1421 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1421 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1421 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1588 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1437 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1437 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1437 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1438 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#371 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#371 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#371 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#371 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#371 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#371 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1588 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1588 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1588 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1621 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1278 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1278 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1621 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1621 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1621 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1622 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1278 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1622 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1622 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1279 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1469 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1469 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1469 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1469 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1470 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_wait.lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1333 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1333 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1333 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1334 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1334 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1334 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1477 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1477 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1477 irq_context: 0 (wq_completion)wg-crypt-wg0#21 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1478 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1478 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1478 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1478 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1334 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1334 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1334 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1334 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1334 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1335 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1335 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1335 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1335 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1336 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1336 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1479 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1479 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1480 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1480 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1480 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1336 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1336 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1337 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1337 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1338 irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1481 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1481 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1482 irq_context: 0 (wq_completion)wg-crypt-wg0#20 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1482 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1338 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1338 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1339 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1483 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1483 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1339 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1339 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1339 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1339 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1340 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1483 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1484 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1484 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1340 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1340 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1340 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1340 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1340 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1341 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1341 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1341 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1341 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1341 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1341 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1342 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#360 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#360 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#360 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#360 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#360 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#360 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#258 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#46 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#46 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#258 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#258 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#46 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#35 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#46 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#46 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#46 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#35 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1485 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1486 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1486 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1342 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1342 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1342 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1342 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1342 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1342 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1342 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1343 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1486 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1486 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1486 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1343 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1343 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1344 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1488 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1488 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1488 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1488 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1344 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1344 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1344 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1344 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1344 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1345 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1489 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1489 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1489 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1345 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1345 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1345 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1345 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1346 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1346 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1347 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1347 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1347 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1347 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1348 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1348 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1348 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1349 irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#44 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1349 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1349 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1350 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1350 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1350 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1350 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1351 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1494 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1495 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1495 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1495 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1351 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1351 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1351 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1352 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1496 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1496 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1352 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1352 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1352 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1352 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1353 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1497 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1497 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1497 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1353 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1353 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1354 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#361 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#361 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1497 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1497 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1498 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1498 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1498 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1498 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1354 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1354 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1354 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1354 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1355 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1498 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1498 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1499 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1499 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1355 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1356 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1501 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1501 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1356 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1356 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1356 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1356 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1356 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1357 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1357 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1504 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1504 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1504 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1359 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1359 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1359 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1360 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1360 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1360 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1360 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1360 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1361 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1506 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1506 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1361 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1361 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1361 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1362 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 tty_mutex &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1363 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1363 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1507 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1507 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1507 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1507 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#362 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#259 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#362 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#362 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#259 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#259 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#362 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#259 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#362 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#362 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#362 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#259 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#259 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#362 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1363 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1363 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1363 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1363 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1363 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1364 irq_context: 0 tty_mutex kfence_freelist_lock irq_context: 0 tty_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1508 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1508 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1509 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1509 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1509 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1364 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1364 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1364 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1364 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1365 irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1365 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1365 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1366 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1366 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1366 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1366 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1366 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1512 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1512 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1512 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1367 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1367 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1367 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1367 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1368 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1368 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1368 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1369 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1369 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1369 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1370 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1515 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1515 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1370 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1370 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1370 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1370 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1370 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1371 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#363 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#363 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#363 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#363 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1515 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1515 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1516 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1516 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1516 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1516 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1371 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1371 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1372 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#364 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#364 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#364 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#364 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#364 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1517 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1517 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1517 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1517 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1518 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1518 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1372 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1372 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1372 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1372 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1372 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1373 irq_context: 0 &type->i_mutex_dir_key#4 &xa->xa_lock#5 irq_context: 0 &type->i_mutex_dir_key#4 &xa->xa_lock#5 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 stock_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &xa->xa_lock#5 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem stock_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 stock_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#53 &rq->__lock irq_context: 0 kn->active#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#53 fs_reclaim irq_context: 0 kn->active#53 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#53 stock_lock irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1521 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1521 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1374 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1374 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1374 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1374 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1375 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1375 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1375 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1522 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1375 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1375 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1375 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1375 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1376 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1376 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1376 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1376 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1376 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1376 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1376 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1376 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1377 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1377 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1377 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1524 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#365 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1524 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1525 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1525 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#260 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#260 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#260 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#260 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#260 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &rq->__lock &base->lock irq_context: 0 &mm->mmap_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1525 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1526 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1526 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1526 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1526 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1526 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1526 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1379 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1379 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1379 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1379 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1379 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1380 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1380 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1380 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1380 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1380 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1380 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1380 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1380 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1381 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#366 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#366 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#366 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#366 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#366 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1527 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1527 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1527 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1528 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1528 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1528 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1528 irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#3 stock_lock irq_context: 0 sb_writers#3 key irq_context: 0 sb_writers#3 pcpu_lock irq_context: 0 sb_writers#3 percpu_counters_lock irq_context: 0 sb_writers#3 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1528 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1528 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1528 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1529 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1381 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1381 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1381 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1381 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1381 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1382 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1382 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1382 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1529 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1530 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1530 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1382 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1382 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1382 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1382 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1382 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1383 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1531 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1531 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1383 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1383 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1383 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1383 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1383 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1384 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1384 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#367 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#367 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#367 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1531 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1531 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1532 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1532 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1532 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1532 irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1533 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1533 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1533 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1384 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1384 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1384 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1384 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1385 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1385 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1385 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1534 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1534 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1385 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1385 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1385 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1385 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1385 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1386 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1386 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#368 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#368 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#368 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#368 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#368 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#368 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#261 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#261 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#261 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#261 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#261 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#261 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1534 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1535 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1535 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1386 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1386 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1386 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1386 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1386 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1387 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1536 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1387 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1387 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1387 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1388 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1388 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1536 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1536 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1537 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1537 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1537 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1388 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1388 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1388 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1389 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1389 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#9 irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1#9 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex kfence_freelist_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#10 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#10 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1538 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1389 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1389 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1389 irq_context: 0 rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1389 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1390 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1390 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1390 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1390 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1390 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1391 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1540 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1391 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1391 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1392 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#70 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1622 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1622 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1623 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1623 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1623 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#45 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#46 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1541 irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1541 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#46 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1541 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1541 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1392 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1392 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1392 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1392 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1392 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1393 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#23 irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#45 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#45 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1623 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1470 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1470 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1470 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1471 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#374 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#374 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#374 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#374 irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#46 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#374 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#374 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1623 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1623 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1624 irq_context: 0 (wq_completion)wg-kex-wg2#46 irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1541 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1541 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1541 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1542 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1542 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1393 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1393 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1393 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1393 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1393 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1394 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1394 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1542 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1542 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1542 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1543 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1543 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1543 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1543 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1394 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1394 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1394 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1394 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1394 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1395 irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#45 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#46 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1624 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1624 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1624 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1624 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1624 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1625 &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1625 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1625 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1471 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1471 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1471 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1472 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1543 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1543 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1543 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1544 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1544 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1395 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1395 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1395 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1396 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1396 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1544 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1544 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1545 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1545 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1545 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1545 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1396 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1396 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1396 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1397 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1397 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#49 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#49 irq_context: 0 (wq_completion)bond0#23 irq_context: 0 (wq_completion)bond0#23 &rq->__lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1604 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1604 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1605 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&hdev->cmd_work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1625 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1625 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#22 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#50 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1625 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1605 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#54 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1545 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1545 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1546 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1546 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1546 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1397 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1397 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1397 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1398 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#27 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy50 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 kn->active#13 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#23 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 kn->active#14 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#18 irq_context: 0 rtnl_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#18 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1547 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1398 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1398 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1398 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1398 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1398 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1399 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1547 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1399 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1399 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1399 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1400 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1400 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1671 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1672 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1672 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1672 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1672 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1672 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1672 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1672 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1673 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1673 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1673 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1673 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1626 irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy53 irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#34 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#70 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#70 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1453 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1453 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1626 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1472 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1472 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1472 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1472 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1473 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1473 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1473 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1626 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1627 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1627 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1627 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1627 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1473 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1473 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1473 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1474 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1453 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#50 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#49 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1454 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#50 irq_context: 0 (wq_completion)wg-kex-wg0#50 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1548 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1548 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1548 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1548 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1400 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1401 irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#25 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#50 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 pcpu_alloc_mutex &rq->__lock cpu_asid_lock irq_context: 0 pcpu_alloc_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1549 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1549 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1550 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1550 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1550 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1401 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1401 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1401 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1401 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1402 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#50 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1605 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1605 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1605 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1550 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1550 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1551 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1551 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1402 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1402 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#50 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1403 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#25 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#23 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#50 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#49 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#25 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#6 &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1646 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1646 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1646 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1476 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1476 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1476 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1476 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1477 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1477 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1477 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1647 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1647 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1648 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1648 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1648 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1648 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1477 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1477 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1477 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1478 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1648 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1648 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1649 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1649 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1649 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1649 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1650 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1650 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1650 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1650 irq_context: 0 &f->f_pos_lock sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#7 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#7 &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#7 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#7 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 khugepaged_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 khugepaged_wait.lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 khugepaged_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 khugepaged_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 khugepaged_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#7 &of->mutex kn->active#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#7 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1650 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1651 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1651 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1652 irq_context: 0 kn->active#53 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#53 remove_cache_srcu &c->lock irq_context: 0 kn->active#53 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#53 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#53 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#53 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#53 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#53 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1652 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1652 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1652 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1652 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1653 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1653 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1653 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1653 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1654 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1654 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1654 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1654 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1478 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1478 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1478 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1478 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1478 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1479 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1479 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1479 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1605 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1606 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1606 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1454 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1454 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1454 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1454 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1455 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1455 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 key irq_context: 0 &type->i_mutex_dir_key#4 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#4 percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#4 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1655 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1655 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1657 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1657 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1658 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1658 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1479 irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1479 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1479 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#35 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#25 irq_context: 0 (wq_completion)wg-kex-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#77 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#77 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#78 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#78 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#79 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#79 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#80 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#80 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#81 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#81 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#82 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#82 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#83 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#83 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#64 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#65 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#66 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#67 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#68 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1480 irq_context: 0 (wq_completion)wg-crypt-wg0#30 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1658 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1659 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1659 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1659 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1659 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1480 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1481 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1606 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1606 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1606 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1606 irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1607 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1607 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1607 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1607 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1455 irq_context: 0 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle rcu_node_0 irq_context: 0 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1456 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1673 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1673 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1673 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1673 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1674 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1674 irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1675 irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#372 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#372 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#372 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#55 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#55 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#372 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc4_nci_rx_wq#372 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1666 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1666 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1666 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1667 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1667 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1667 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1667 irq_context: 0 (wq_completion)wg-kex-wg2#56 irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 rtnl_mutex stack_depot_init_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#372 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1607 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1607 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1607 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond0#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1608 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1608 irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1667 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1667 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1667 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#78 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#79 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#80 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#81 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#82 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&tn->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&tn->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#83 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#83 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 key#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1608 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1608 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1608 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#23 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1608 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1609 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1609 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#20 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#20 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1611 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1458 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1458 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1458 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1459 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1667 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1668 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1668 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#12 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)hci4#12 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#15 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#20 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci2#11 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1611 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1611 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1611 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1612 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1612 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1612 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1459 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1459 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1459 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1459 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1460 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1460 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1460 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#373 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#373 &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#373 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#373 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#373 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#373 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#373 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#373 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#262 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#262 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#262 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#262 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) pool_lock#2 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#11 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1612 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1612 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1612 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1613 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1613 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1460 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1460 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1460 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1461 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#12 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#12 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#12 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 rtnl_mutex remove_cache_srcu &base->lock irq_context: 0 rtnl_mutex remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: hardirq rcu_state.barrier_lock fill_pool_map-wait-type-override &c->lock irq_context: hardirq rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#57 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#60 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#58 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1793 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1793 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1793 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1789 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1552 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1552 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1552 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1551 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1551 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#423 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1613 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1613 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1614 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1614 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#423 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#423 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#423 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#423 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#423 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#75 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#75 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#75 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1461 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1461 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1461 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1461 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1462 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#75 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#75 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#56 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#56 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#56 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#56 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#56 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#56 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1794 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1794 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1794 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1794 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1790 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1553 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1553 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1553 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1552 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1552 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#424 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#424 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#424 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#424 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#424 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &meta->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1795 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1795 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1795 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1795 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1791 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1791 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1791 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1554 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1554 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1554 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1554 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1553 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1553 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1553 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#425 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#425 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#425 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#425 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#425 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1792 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1792 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1792 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1555 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1554 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1554 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1555 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1555 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1555 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#292 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#292 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#292 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#426 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#292 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#426 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1797 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1797 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#426 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#426 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1797 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#426 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#426 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1793 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1793 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1793 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#76 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#76 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#76 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#76 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1798 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1798 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1794 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1794 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1557 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1557 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1557 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1799 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1799 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1799 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1795 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1800 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1677 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1677 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1677 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#47 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1677 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1678 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1678 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1678 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1678 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock quarantine_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#263 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#263 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#263 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#263 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#263 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#263 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#47 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#47 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#47 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#47 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#47 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#36 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#36 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#36 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#36 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#7 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq irq_context: 0 (wq_completion)nfc10_nci_rx_wq irq_context: 0 (wq_completion)nfc10_nci_rx_wq &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq irq_context: 0 (wq_completion)nfc10_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq irq_context: 0 (wq_completion)nfc11_nci_rx_wq irq_context: 0 (wq_completion)nfc11_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq irq_context: 0 (wq_completion)nfc12_nci_rx_wq irq_context: 0 (wq_completion)nfc12_nci_tx_wq irq_context: 0 (wq_completion)nfc12_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq irq_context: 0 (wq_completion)nfc13_nci_cmd_wq irq_context: 0 (wq_completion)nfc13_nci_rx_wq irq_context: 0 (wq_completion)nfc13_nci_tx_wq irq_context: 0 (wq_completion)nfc14_nci_rx_wq irq_context: 0 (wq_completion)nfc14_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_cmd_wq irq_context: 0 (wq_completion)nfc15_nci_rx_wq irq_context: 0 (wq_completion)nfc15_nci_tx_wq irq_context: 0 (wq_completion)nfc15_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq irq_context: 0 (wq_completion)nfc16_nci_cmd_wq irq_context: 0 (wq_completion)nfc16_nci_rx_wq irq_context: 0 (wq_completion)nfc16_nci_tx_wq irq_context: 0 (wq_completion)nfc16_nci_tx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq irq_context: 0 (wq_completion)nfc17_nci_cmd_wq irq_context: 0 (wq_completion)nfc17_nci_rx_wq irq_context: 0 (wq_completion)nfc17_nci_tx_wq irq_context: 0 (wq_completion)nfc17_nci_tx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nfc18_nci_cmd_wq irq_context: 0 (wq_completion)nfc18_nci_rx_wq irq_context: 0 (wq_completion)nfc18_nci_rx_wq &rq->__lock irq_context: 0 (wq_completion)nfc18_nci_rx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc18_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq irq_context: 0 (wq_completion)nfc19_nci_cmd_wq irq_context: 0 (wq_completion)nfc19_nci_rx_wq irq_context: 0 (wq_completion)nfc19_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq irq_context: 0 (wq_completion)nfc21_nci_cmd_wq irq_context: 0 (wq_completion)nfc21_nci_rx_wq irq_context: 0 (wq_completion)nfc21_nci_rx_wq &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_rx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc21_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq irq_context: 0 (wq_completion)nfc22_nci_cmd_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 (wq_completion)nfc22_nci_rx_wq irq_context: 0 (wq_completion)nfc22_nci_tx_wq irq_context: 0 (wq_completion)nfc22_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc22_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq irq_context: 0 (wq_completion)nfc23_nci_cmd_wq irq_context: 0 (wq_completion)nfc23_nci_rx_wq irq_context: 0 (wq_completion)nfc23_nci_tx_wq irq_context: 0 (wq_completion)nfc23_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc23_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_cmd_wq irq_context: 0 (wq_completion)nfc24_nci_rx_wq irq_context: 0 (wq_completion)nfc24_nci_rx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq irq_context: 0 (wq_completion)nfc25_nci_cmd_wq irq_context: 0 (wq_completion)nfc25_nci_rx_wq irq_context: 0 (wq_completion)nfc25_nci_tx_wq irq_context: 0 (wq_completion)nfc25_nci_tx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_cmd_wq irq_context: 0 (wq_completion)nfc26_nci_rx_wq irq_context: 0 (wq_completion)nfc26_nci_tx_wq irq_context: 0 (wq_completion)nfc26_nci_tx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_cmd_wq irq_context: 0 (wq_completion)nfc27_nci_rx_wq irq_context: 0 (wq_completion)nfc27_nci_tx_wq irq_context: 0 (wq_completion)nfc27_nci_tx_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1680 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1680 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1680 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1681 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1681 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1681 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1484 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1484 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1484 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1484 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1485 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#376 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#376 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#376 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#376 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#376 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#264 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#264 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#264 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#264 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#264 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#48 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#48 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#48 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#37 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#37 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#37 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#37 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#37 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#8 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#8 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#8 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#3 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#3 irq_context: 0 (wq_completion)wg-crypt-wg0#24 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#3 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#3 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#2 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#2 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#2 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#2 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#2 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc56_nci_cmd_wq irq_context: 0 (wq_completion)nfc56_nci_cmd_wq irq_context: 0 (wq_completion)nfc56_nci_rx_wq irq_context: 0 (wq_completion)nfc56_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc55_nci_cmd_wq irq_context: 0 (wq_completion)nfc55_nci_cmd_wq irq_context: 0 (wq_completion)nfc55_nci_rx_wq irq_context: 0 (wq_completion)nfc55_nci_tx_wq irq_context: 0 (wq_completion)nfc55_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc55_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc54_nci_cmd_wq irq_context: 0 (wq_completion)nfc54_nci_rx_wq irq_context: 0 (wq_completion)nfc54_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc53_nci_cmd_wq irq_context: 0 (wq_completion)nfc53_nci_cmd_wq irq_context: 0 (wq_completion)nfc53_nci_rx_wq irq_context: 0 (wq_completion)nfc53_nci_tx_wq irq_context: 0 (wq_completion)nfc53_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc53_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc52_nci_cmd_wq irq_context: 0 (wq_completion)nfc52_nci_cmd_wq irq_context: 0 (wq_completion)nfc52_nci_rx_wq irq_context: 0 (wq_completion)nfc52_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc51_nci_cmd_wq irq_context: 0 (wq_completion)nfc51_nci_cmd_wq irq_context: 0 (wq_completion)nfc51_nci_rx_wq irq_context: 0 (wq_completion)nfc51_nci_tx_wq irq_context: 0 &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc50_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc50_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc50_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc50_nci_cmd_wq irq_context: 0 (wq_completion)nfc50_nci_rx_wq irq_context: 0 (wq_completion)nfc50_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc49_nci_cmd_wq irq_context: 0 (wq_completion)nfc49_nci_cmd_wq irq_context: 0 (wq_completion)nfc49_nci_rx_wq irq_context: 0 (wq_completion)nfc49_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc48_nci_cmd_wq irq_context: 0 (wq_completion)nfc48_nci_cmd_wq irq_context: 0 (wq_completion)nfc48_nci_rx_wq irq_context: 0 (wq_completion)nfc48_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc47_nci_cmd_wq irq_context: 0 (wq_completion)nfc47_nci_cmd_wq irq_context: 0 (wq_completion)nfc47_nci_rx_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#21 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#21 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &n->list_lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)nfc46_nci_cmd_wq irq_context: 0 (wq_completion)nfc46_nci_rx_wq irq_context: 0 (wq_completion)nfc46_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc45_nci_cmd_wq irq_context: 0 (wq_completion)nfc45_nci_cmd_wq irq_context: 0 (wq_completion)nfc45_nci_rx_wq irq_context: 0 (wq_completion)nfc45_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc44_nci_cmd_wq irq_context: 0 (wq_completion)nfc44_nci_cmd_wq irq_context: 0 (wq_completion)nfc44_nci_rx_wq irq_context: 0 (wq_completion)nfc44_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc43_nci_cmd_wq irq_context: 0 (wq_completion)nfc43_nci_cmd_wq irq_context: 0 (wq_completion)nfc43_nci_rx_wq irq_context: 0 (wq_completion)nfc43_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq irq_context: 0 (wq_completion)nfc42_nci_cmd_wq irq_context: 0 (wq_completion)nfc42_nci_rx_wq irq_context: 0 (wq_completion)nfc42_nci_tx_wq irq_context: 0 (wq_completion)nfc42_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc42_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc41_nci_cmd_wq irq_context: 0 (wq_completion)nfc41_nci_rx_wq irq_context: 0 (wq_completion)nfc41_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq irq_context: 0 (wq_completion)nfc40_nci_cmd_wq irq_context: 0 (wq_completion)nfc40_nci_rx_wq irq_context: 0 (wq_completion)nfc40_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq irq_context: 0 wq_pool_attach_mutex rcu_node_0 irq_context: 0 wq_pool_attach_mutex &rcu_state.expedited_wq irq_context: 0 wq_pool_attach_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 wq_pool_attach_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 wq_pool_attach_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc39_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1668 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1668 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1668 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1669 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1669 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq irq_context: 0 (wq_completion)nfc38_nci_cmd_wq irq_context: 0 (wq_completion)nfc38_nci_rx_wq irq_context: 0 (wq_completion)nfc38_nci_tx_wq irq_context: 0 (wq_completion)nfc38_nci_tx_wq &rq->__lock irq_context: 0 (wq_completion)nfc38_nci_tx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq irq_context: 0 (wq_completion)nfc37_nci_cmd_wq irq_context: 0 (wq_completion)nfc37_nci_rx_wq irq_context: 0 (wq_completion)nfc37_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq irq_context: 0 (wq_completion)nfc36_nci_cmd_wq irq_context: 0 (wq_completion)nfc36_nci_rx_wq irq_context: 0 (wq_completion)nfc36_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_cmd_wq irq_context: 0 (wq_completion)bond0#27 irq_context: 0 (wq_completion)bond0#27 &rq->__lock irq_context: 0 (wq_completion)bond0#27 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#27 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc35_nci_rx_wq irq_context: 0 (wq_completion)nfc35_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq irq_context: 0 (wq_completion)nfc34_nci_cmd_wq irq_context: 0 (wq_completion)nfc34_nci_rx_wq irq_context: 0 (wq_completion)nfc34_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq irq_context: 0 (wq_completion)nfc33_nci_cmd_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1669 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1669 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#28 irq_context: 0 (wq_completion)wg-crypt-wg0#28 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1670 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1670 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1670 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1670 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1670 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1670 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex pcpu_alloc_mutex fs_reclaim irq_context: 0 rtnl_mutex pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)nfc33_nci_rx_wq irq_context: 0 (wq_completion)wg-kex-wg1#62 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci1#12 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy50 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#62 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#61 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#61 irq_context: 0 (wq_completion)wg-kex-wg2#61 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#62 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc33_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq irq_context: 0 (wq_completion)nfc32_nci_cmd_wq irq_context: 0 (wq_completion)nfc32_nci_rx_wq irq_context: 0 (wq_completion)nfc32_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq irq_context: 0 (wq_completion)nfc31_nci_cmd_wq irq_context: 0 (wq_completion)nfc31_nci_rx_wq irq_context: 0 (wq_completion)nfc31_nci_tx_wq irq_context: 0 (wq_completion)nfc30_nci_cmd_wq irq_context: 0 (wq_completion)nfc30_nci_rx_wq irq_context: 0 (wq_completion)nfc30_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq irq_context: 0 (wq_completion)nfc29_nci_cmd_wq irq_context: 0 (wq_completion)nfc29_nci_rx_wq irq_context: 0 (wq_completion)nfc29_nci_rx_wq &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_rx_wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq irq_context: 0 (wq_completion)nfc28_nci_cmd_wq irq_context: 0 (wq_completion)nfc28_nci_rx_wq irq_context: 0 (wq_completion)nfc28_nci_tx_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1485 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1485 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1485 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1486 irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1687 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1686 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1686 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1686 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock pcpu_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock irq_context: 0 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 rcu_read_lock &____s->seqcount#4 irq_context: 0 sb_writers#3 &zone->lock irq_context: 0 sb_writers#3 &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &p->alloc_lock irq_context: 0 sb_writers#3 rcu_read_lock &____s->seqcount#3 irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &f->f_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &xa->xa_lock#5 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock percpu_counters_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#11 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#11 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &retval->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &retval->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 stock_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1688 irq_context: 0 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1688 &rq->__lock irq_context: 0 &ei->i_data_sem pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1688 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#7 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1688 irq_context: 0 sb_writers#3 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1688 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1688 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1688 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1687 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#3 &mm->mmap_lock key irq_context: 0 sb_writers#3 &mm->mmap_lock pcpu_lock irq_context: 0 sb_writers#3 &mm->mmap_lock percpu_counters_lock irq_context: 0 sb_writers#3 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#32 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_internal jbd2_handle &ei->i_prealloc_lock irq_context: 0 sb_internal jbd2_handle &journal->j_list_lock irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#16 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1688 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1688 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1690 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1690 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1690 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1690 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1690 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1689 irq_context: 0 sb_writers#3 &mm->mmap_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock key irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1691 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1691 irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1690 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 sb_writers#3 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &____s->seqcount#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1692 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1692 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1692 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1691 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock key#9 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#51 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#53 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#54 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#51 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#52 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#27 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1693 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1693 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1693 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1693 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1693 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1693 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1693 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1692 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1692 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1692 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 pcpu_lock stock_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem &meta->lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1694 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1694 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1694 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1694 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1693 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1693 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1693 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1695 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1695 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1695 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1695 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1694 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &pl->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &pl->lock key#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex minors_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&uhid->worker) &dev->mutex minors_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1696 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1696 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &sem->wait_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override &__ctx->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &tags->lock irq_context: 0 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#274 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#274 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#274 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1729 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1729 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1729 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1729 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1728 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#59 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#59 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#59 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#59 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#59 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#59 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#393 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#393 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#393 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#393 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#393 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#393 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#62 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1510 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1510 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1510 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1510 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1510 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1510 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1510 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1510 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1510 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1510 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1511 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1511 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1511 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#395 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#395 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#395 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#395 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#395 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#395 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#276 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#49 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#276 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#276 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#49 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#49 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#276 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#49 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#49 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#278 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#278 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1515 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1515 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1515 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1515 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1515 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#279 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#279 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#279 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#279 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#279 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#279 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#279 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#279 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#398 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#398 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#398 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#398 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1733 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1733 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1733 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#280 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#280 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#280 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1733 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1733 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1733 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1732 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#65 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#65 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#65 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1516 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1516 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1516 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1516 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1516 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#51 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#51 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#51 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#51 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#51 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#51 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1517 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1517 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1517 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1517 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1517 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#399 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#399 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#399 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#399 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#399 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#400 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#400 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1518 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1518 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#400 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#281 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#281 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#281 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1518 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#281 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#281 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#281 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1518 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1518 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#281 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#52 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#52 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#281 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1518 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#7 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1734 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1734 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1734 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1734 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1734 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1734 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1733 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#66 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#66 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#66 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#401 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#401 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#401 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#401 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#401 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#53 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#282 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#53 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#282 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#282 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#282 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#53 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#53 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#53 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#53 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#67 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#67 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#67 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1519 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1519 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1519 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1519 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1735 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1735 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1735 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1735 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1735 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1734 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#402 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#402 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#402 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#402 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#68 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#68 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#68 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#68 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#68 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#68 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#403 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#403 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#403 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1520 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1520 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1736 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1736 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1736 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1735 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#404 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#404 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#69 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1737 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1737 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1737 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1737 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1737 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#69 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1736 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1736 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1736 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#405 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc4_nci_rx_wq#405 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#405 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#405 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#405 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1738 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1738 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1738 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1738 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1738 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1737 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1737 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1521 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1737 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1521 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1521 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#284 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#284 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#284 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#284 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#284 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#284 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1739 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1739 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1739 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1739 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1739 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1739 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1738 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#406 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#406 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#406 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#406 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#406 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#406 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1522 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1522 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1522 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1522 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1523 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1523 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#407 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#407 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#407 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#407 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#407 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#407 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#407 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#407 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#408 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#408 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#285 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#285 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#285 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#285 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#409 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#409 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#70 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#409 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#409 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1524 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1524 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1524 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1524 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1524 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1740 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1740 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1740 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1740 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1740 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1740 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1740 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1739 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1739 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1739 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1741 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1740 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1742 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1742 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1742 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1742 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1741 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1743 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1742 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#410 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#410 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#410 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1744 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1744 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1744 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1744 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1744 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1743 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1743 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1743 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1745 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1745 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1527 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1527 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1527 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1527 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1527 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1744 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1746 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1745 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1528 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1528 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1528 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1528 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1528 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1528 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1528 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1529 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1529 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1529 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1529 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1748 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1747 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#415 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#415 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#415 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#415 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#415 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1537 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1537 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1537 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1757 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1757 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1757 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1757 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1757 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1756 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1539 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1539 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1758 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1758 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1758 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1758 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1758 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1758 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1757 irq_context: 0 (wq_completion)wg-kex-wg1#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#56 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1759 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem key irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1760 pool_lock#2 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1760 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1760 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 pernet_ops_rwsem &nft_net->commit_mutex irq_context: 0 pernet_ops_rwsem netns_bpf_mutex irq_context: 0 pernet_ops_rwsem &x->wait#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1766 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1766 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1763 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1544 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1544 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1544 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1758 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1544 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1544 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#289 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#289 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#289 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex class irq_context: 0 pernet_ops_rwsem rtnl_mutex (&tbl->proxy_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ul->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->xdp.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex mirred_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ent->pde_unload_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_report_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnn->routes.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex target_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL6 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#289 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#73 irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex deferred_probe_mutex irq_context: 0 (wq_completion)nfc6_nci_rx_wq#73 irq_context: 0 pernet_ops_rwsem rtnl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#420 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#420 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#420 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#73 irq_context: 0 pernet_ops_rwsem rtnl_mutex device_links_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#420 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &x->wait#24 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#420 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#420 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#420 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#420 &rq->__lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 pool_lock#2 irq_context: 0 pernet_ops_rwsem netdev_unregistering_wq.lock irq_context: 0 pernet_ops_rwsem (work_completion)(&ht->run_work) irq_context: 0 pernet_ops_rwsem &ht->mutex irq_context: 0 pernet_ops_rwsem &ht->mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_ETHER irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1767 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1767 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1767 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1764 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1764 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1545 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1545 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1545 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1545 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1545 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1545 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#421 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#421 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#421 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#421 irq_context: 0 pernet_ops_rwsem quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#421 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#421 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &ul->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem rdma_nets.xa_lock irq_context: 0 pernet_ops_rwsem k-clock-AF_NETLINK irq_context: 0 pernet_ops_rwsem &nlk->wait irq_context: 0 pernet_ops_rwsem &xa->xa_lock#9 irq_context: 0 pernet_ops_rwsem &fsnotify_mark_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_LOOPBACK irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1768 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1768 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1768 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1765 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1765 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1765 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1546 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1546 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1546 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1546 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &hn->hn_lock irq_context: 0 pernet_ops_rwsem sysctl_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem sysctl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem sysctl_lock krc.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &x->wait#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1546 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#422 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#422 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#422 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#290 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#290 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#290 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#290 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#74 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#74 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#74 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#74 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &this->info_list_lock irq_context: 0 pernet_ops_rwsem &pnettable->lock irq_context: 0 (wq_completion)bond0#28 irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#37 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#37 irq_context: 0 pernet_ops_rwsem &pnetids_ndev->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 &list->lock#19 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#35 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#104 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#104 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#104 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1671 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1671 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#342 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#342 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#342 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#342 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#341 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#341 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#341 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#341 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#503 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#503 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#501 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#501 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1673 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1673 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1673 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1673 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1672 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1672 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1672 irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#504 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#504 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#504 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#502 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#502 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1674 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1674 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1674 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1674 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1673 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1673 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#343 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#343 &rq->__lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#74 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 pool_lock#2 irq_context: 0 &ep->mtx &mm->mmap_lock rcu_node_0 irq_context: 0 &ep->mtx &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &ep->mtx &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ep->mtx &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#74 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1769 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1769 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1769 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1769 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1769 irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#28 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1766 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#55 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#55 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#55 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#55 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#55 irq_context: 0 rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#55 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1547 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1547 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1547 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1770 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1770 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1770 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1770 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1767 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#343 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#106 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#106 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#343 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#342 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#342 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#106 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#105 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1958 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#105 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#105 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1958 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1958 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock irq_context: 0 pernet_ops_rwsem ip6_fl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->rules_mod_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#28 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#28 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#28 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem (&net->ipv6.ip6_fib_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex (&mrt->ipmr_expire_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#28 irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#29 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#29 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#31 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#31 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#32 irq_context: 0 pernet_ops_rwsem (&ipvs->dest_trash_timer) irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#32 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) irq_context: 0 pernet_ops_rwsem nfnl_subsys_ipset irq_context: 0 pernet_ops_rwsem recent_lock irq_context: 0 pernet_ops_rwsem hashlimit_mutex irq_context: 0 pernet_ops_rwsem trans_gc_work irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock irq_context: 0 pernet_ops_rwsem tcp_metrics_lock irq_context: 0 pernet_ops_rwsem k-clock-AF_INET irq_context: 0 pernet_ops_rwsem (work_completion)(&net->xfrm.policy_hash_work) irq_context: 0 pernet_ops_rwsem &net->xfrm.xfrm_policy_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&net->xfrm.state_hash_work) irq_context: 0 pernet_ops_rwsem &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem genl_sk_destructing_waitq.lock irq_context: 0 (wq_completion)events fqdir_free_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &list->lock#19 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_ep_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 clock-AF_INET irq_context: softirq rcu_callback &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#25 irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#26 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#26 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#27 pool_lock#2 irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#28 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#29 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#30 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#31 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#32 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#32 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 kn->active#11 remove_cache_srcu irq_context: 0 kn->active#11 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#11 remove_cache_srcu &c->lock irq_context: 0 kn->active#11 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#11 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#11 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#11 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-crypt-wg0#27 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#26 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1771 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1771 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1771 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1771 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1771 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1771 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1768 irq_context: 0 pernet_ops_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &base->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &____s->seqcount#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1769 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1773 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1773 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&app->join_timer) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#23 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#23 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#69 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#23 irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#70 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#35 irq_context: 0 (wq_completion)wg-crypt-wg0#35 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1800 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#427 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#427 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#427 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#427 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1796 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1796 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1558 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1558 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1558 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1558 irq_context: 0 rtnl_mutex team->team_lock_key#33 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#33 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#33 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#33 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1557 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1557 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1801 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1801 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1801 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1801 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1797 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1798 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1564 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1564 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1563 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1563 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1563 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1815 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1811 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1565 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1565 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1564 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1564 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1566 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1566 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1565 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1565 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#23 &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1565 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1817 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1566 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1566 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1566 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1818 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1818 irq_context: 0 rtnl_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1818 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1818 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1818 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1814 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1814 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1814 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1568 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1568 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1568 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1567 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1567 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &rq->__lock irq_context: 0 (wq_completion)hci4#9 irq_context: 0 (wq_completion)hci4#9 &rq->__lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1569 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1568 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1568 &rq->__lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#20 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#31 irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#35 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#62 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy54 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1568 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#84 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#84 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#85 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#85 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#86 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#86 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#86 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#87 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#87 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#87 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#88 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#88 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#88 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#88 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#89 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#89 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) rcu_node_0 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#90 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#90 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#90 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#91 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#91 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#91 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#92 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#92 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#93 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#93 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#94 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#94 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#95 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#95 irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#96 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#96 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#97 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#97 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#9 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &____s->seqcount irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#10 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#10 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#10 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#98 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#98 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#99 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#99 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#99 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#100 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#100 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#101 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#101 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#102 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#102 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#102 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#102 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#102 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#102 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#103 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#103 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#104 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#104 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#105 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#105 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#105 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#106 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#106 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#107 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#107 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#108 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#108 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#108 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#108 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1820 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1820 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1820 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#109 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#109 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#110 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#110 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#111 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#111 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#112 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#112 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#112 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1820 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1820 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1816 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#84 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1570 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#84 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#84 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1570 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1570 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1570 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1569 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1569 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1569 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#86 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#86 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#87 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#87 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#87 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#89 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#89 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#89 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#90 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#91 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#91 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#91 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1571 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1571 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1571 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1571 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1570 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1570 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1570 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1570 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#93 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#94 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#95 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1818 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1572 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1572 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1571 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1571 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1571 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1571 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1819 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1819 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1819 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1824 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1820 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#101 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#102 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1573 &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1825 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1825 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1825 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1825 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1825 irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1573 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1825 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1825 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1825 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1572 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1821 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1821 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1821 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1572 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1822 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1822 &rq->__lock irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#29 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1822 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#25 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1574 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1574 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1574 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1573 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1573 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1573 irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1827 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1827 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1827 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1823 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1828 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1828 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1828 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1828 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1828 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1824 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1824 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1575 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1575 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1575 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1575 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1574 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1574 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#107 irq_context: 0 (wq_completion)wg-crypt-wg1#28 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1829 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1829 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1829 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#432 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#432 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#108 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1825 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#109 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#110 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#110 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#432 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#432 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#111 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#112 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1576 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1576 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1576 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1575 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1575 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1830 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1830 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1830 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1830 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1830 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1826 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#433 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#433 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#433 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#433 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#433 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#433 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#295 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#295 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#295 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#295 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#295 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#295 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#295 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#295 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1831 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1831 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1831 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1831 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1827 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1577 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1577 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1576 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1576 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1576 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1576 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1576 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#434 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#434 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#434 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#434 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#434 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#434 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1832 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1832 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1832 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1832 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1828 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1828 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#296 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#296 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#296 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#296 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#296 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#296 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#296 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &mapping->i_private_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1578 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1578 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1578 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1578 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1578 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1577 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1577 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1833 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1833 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1833 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sem->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1579 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1833 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1833 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1833 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1829 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#77 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#77 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1579 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1578 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1578 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#77 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#77 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#77 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#435 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#435 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#435 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#435 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#435 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1834 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1834 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1834 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1834 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1834 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1830 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#436 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#436 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1580 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#436 irq_context: 0 (wq_completion)wg-kex-wg2#60 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1579 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1579 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1579 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#436 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1579 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#59 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#30 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1835 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1835 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1835 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1835 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1835 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1831 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#297 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#297 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#297 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#297 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#78 irq_context: 0 sb_writers#3 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1832 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1832 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1832 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1581 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1581 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1581 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1581 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1580 irq_context: 0 pernet_ops_rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1580 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#79 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#79 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#79 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#79 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#79 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#79 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#437 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#437 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#437 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#437 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#437 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#57 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#57 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1837 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1833 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#16 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&conn->pending_rx_work) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override &__ctx->lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1582 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1582 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1582 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1582 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1581 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1581 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1581 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1581 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1581 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1581 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#438 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#438 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1838 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1838 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1834 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#299 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#299 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#299 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#299 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1583 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1583 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1582 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1582 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1594 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1593 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1593 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1593 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#309 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#62 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#309 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#62 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#308 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#308 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#62 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#62 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1595 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1595 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1594 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1594 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#118 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1594 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#450 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#450 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#450 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#450 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#450 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#450 irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1596 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1596 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1596 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1595 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1595 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#86 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#86 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#86 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#86 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1849 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1849 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1849 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1849 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1849 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1845 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1845 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1845 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1597 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1597 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1596 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1596 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1850 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1850 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1850 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1850 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1846 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1598 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1598 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1597 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1597 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#310 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#310 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1597 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#451 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#451 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#309 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#309 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#309 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#309 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#33 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#33 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#33 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#34 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#34 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#34 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#34 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#35 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#35 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#35 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#36 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#36 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#451 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#36 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#36 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#37 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#37 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#37 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#37 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#38 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#38 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#38 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#451 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#39 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#451 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#451 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#39 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#451 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#451 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#451 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#41 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#41 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#41 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#42 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#42 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#43 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#43 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#44 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#44 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#45 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#45 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#46 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#46 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#46 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#46 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#47 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#47 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#47 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#47 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#48 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#49 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#49 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#50 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#50 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#51 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#51 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#52 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#52 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#54 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#54 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#55 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#55 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#56 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#56 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#58 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#58 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#33 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1598 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1598 irq_context: 0 (wq_completion)bond0#30 irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1598 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1598 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#34 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#452 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#452 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#35 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#452 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#30 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &base->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#87 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#87 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#87 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#87 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1600 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1600 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1599 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1599 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1599 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#311 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#311 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#311 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#311 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#310 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#37 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#310 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#310 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#310 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1851 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#38 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#39 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#41 irq_context: 0 rtnl_mutex team->team_lock_key#33 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#33 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 lock irq_context: 0 rtnl_mutex team->team_lock_key#33 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#33 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#33 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &____s->seqcount#2 irq_context: 0 rtnl_mutex team->team_lock_key#33 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#33 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#33 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#33 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#33 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#33 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#33 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#33 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#33 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#42 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#43 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#34 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#44 irq_context: 0 rtnl_mutex team->team_lock_key#34 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#34 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#34 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#34 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#34 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#45 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#46 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#47 irq_context: 0 rtnl_mutex team->team_lock_key#34 lock irq_context: 0 rtnl_mutex team->team_lock_key#34 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#34 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#34 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1851 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#453 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#453 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1851 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1851 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#48 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#49 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1851 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1847 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#453 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#453 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#56 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#57 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#58 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc4_nci_tx_wq#453 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#71 irq_context: 0 tasklist_lock &sighand->siglock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#36 irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#36 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mount_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#35 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#72 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 irq_context: 0 (wq_completion)wg-kex-wg0#65 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#453 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#63 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#63 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1852 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1852 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1852 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1852 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1848 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1853 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1853 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1853 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1853 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1853 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1853 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1853 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1849 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#454 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#454 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#454 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#454 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#454 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1601 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#123 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#123 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#123 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1600 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1600 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1600 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1600 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_tx_wq#25 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_tx_wq#25 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc8_nci_tx_wq#25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#26 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#26 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#26 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#27 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#27 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#64 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#64 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#88 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#88 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#88 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#88 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#88 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#312 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#312 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#311 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#311 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1602 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1602 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1601 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1601 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#455 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#455 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#455 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#455 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1854 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1854 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1854 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1854 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1850 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1855 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1855 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1851 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#65 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#65 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#65 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#127 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#127 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#28 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#28 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#28 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#28 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#28 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#26 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#8 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#8 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#8 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#456 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#456 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#456 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1602 irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#128 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#15 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#128 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#128 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#89 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#89 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#89 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#89 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#89 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#89 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#89 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#89 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#132 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#132 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1857 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1857 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1857 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1857 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1857 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1857 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1857 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1853 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#458 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#458 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#458 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#314 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#458 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#458 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#458 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1604 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1604 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1604 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1858 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1854 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#459 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#459 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#459 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#459 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#459 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#459 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#459 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1606 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1605 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1605 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#316 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#316 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#136 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#90 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#90 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#315 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#315 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#315 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#315 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#315 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1855 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#138 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#460 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#460 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#460 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#460 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#460 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#460 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1606 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1606 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1606 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1606 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1606 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1607 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1607 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1607 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1607 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1609 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1608 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1608 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1608 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1608 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1861 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1861 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1861 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1861 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1861 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1861 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1857 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1857 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1857 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#317 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1610 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1610 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1609 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1609 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1609 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1609 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1609 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc4_nci_tx_wq#462 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#91 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#91 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#91 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#91 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1612 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1612 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1611 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1611 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1864 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1864 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1860 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1860 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1613 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1613 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1612 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1612 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1612 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#319 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#319 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#463 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#463 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#463 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#463 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#463 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#463 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#463 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#318 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#318 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1614 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1614 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1614 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1614 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1613 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1613 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1613 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1866 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1862 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1862 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1862 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1615 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1615 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1615 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1615 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1614 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1614 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1614 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#464 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#464 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#464 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#464 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#464 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1867 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1863 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1616 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1616 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1616 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1616 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1615 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1615 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1615 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1615 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#465 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#465 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#465 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#29 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#29 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#465 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#29 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#29 rcu_node_0 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#29 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#465 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#27 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#465 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#27 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#465 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#465 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#465 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#465 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#320 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#320 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#320 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#319 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#319 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#319 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#319 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#92 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#92 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (wq_completion)hci1#12 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#92 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#92 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#92 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#67 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#67 irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1865 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1865 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1865 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#30 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)hci1#13 irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 &rq->__lock irq_context: 0 (wq_completion)hci1#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1870 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1870 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1870 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1870 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#13 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#14 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#14 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1870 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1866 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1866 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1866 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1616 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1871 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1871 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1867 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1867 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1867 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#18 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1872 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1872 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1868 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1873 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1873 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1873 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1873 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1873 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1869 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1869 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1869 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1874 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1874 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1874 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1874 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1874 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1870 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1875 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1871 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1871 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1871 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1873 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1874 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1874 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#35 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#36 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#36 irq_context: 0 (wq_completion)wg-crypt-wg2#36 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#71 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#35 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#36 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#72 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#14 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1879 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1875 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1875 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#62 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#62 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#61 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 irq_context: 0 (wq_completion)wg-crypt-wg2#31 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#31 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1885 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1885 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1885 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1881 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#140 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#28 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)hci1#16 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex.wait_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &p->pi_lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3#10 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1880 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1880 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1880 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1880 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1880 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1880 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1880 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1876 irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &sem->wait_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1881 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1881 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1881 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1881 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1881 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1877 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle pcpu_lock stock_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci3#9 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy38 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 kn->active#52 &n->list_lock irq_context: 0 kn->active#52 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#36 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#65 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#66 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#66 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &____s->seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#10 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#10 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci3#10 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#34 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#34 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1902 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1902 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1902 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1900 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1900 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1900 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1896 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1625 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1625 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1624 irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#34 remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#34 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#34 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#34 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#34 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 fs_reclaim irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1624 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 rcu_read_lock &ndev->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci5#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1903 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1903 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1901 irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1897 irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1626 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1626 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1626 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1626 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1625 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1625 irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#32 irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1902 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1902 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1904 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#76 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_rx_wq#76 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#76 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#38 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#38 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1958 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#76 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1954 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#36 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1954 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#36 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1954 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1949 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1949 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1949 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1959 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1955 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1950 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1950 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1950 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1960 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1960 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1960 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1956 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1951 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1951 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1951 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1675 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1675 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1674 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1674 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1674 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1674 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1674 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1674 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1961 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1961 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1957 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1952 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1962 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1958 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1953 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1953 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#505 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#503 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#503 &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) rcu_node_0 irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#503 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#503 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1676 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1676 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1676 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1676 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1675 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1675 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1675 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1675 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1675 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1675 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci5#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci5#6 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci5#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#33 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#34 &____s->seqcount#2 irq_context: 0 rtnl_mutex team->team_lock_key#34 &____s->seqcount irq_context: 0 (wq_completion)bond0#35 irq_context: 0 (wq_completion)bond0#35 &rq->__lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#141 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#141 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#141 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#141 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#35 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#34 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#35 irq_context: 0 rtnl_mutex team->team_lock_key#35 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#35 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#35 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#35 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#35 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#35 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#35 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#35 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#35 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#35 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#35 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#35 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#35 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#35 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#35 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#35 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#35 &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#63 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#63 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#64 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#64 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#32 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#32 irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1905 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1905 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1905 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1905 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1903 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1898 irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#72 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#12 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#63 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#64 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#64 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#32 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &u->iolock &pcp->lock &zone->lock irq_context: 0 &u->iolock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#72 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#72 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#65 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#65 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#66 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#66 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#33 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy37 irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#71 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#61 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#36 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg2#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#35 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci0#12 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#36 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &rq->__lock irq_context: 0 (wq_completion)bond0#37 irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#37 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#76 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#38 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1729 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#545 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#545 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#545 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &wg->device_update_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#76 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#76 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_node_0 irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#61 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#62 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#76 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#75 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#38 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#122 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#122 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#122 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#47 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &type->s_umount_key#47 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#543 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#122 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#540 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#540 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#540 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#122 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &obj_hash[i].lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#121 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#372 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#372 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#368 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#368 irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#38 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2023 irq_context: 0 &hdev->req_lock (wq_completion)hci0#10 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#33 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2023 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2023 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2023 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2018 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2018 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2018 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2013 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2019 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2014 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1733 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1733 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1733 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1733 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1731 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1731 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#5 irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#36 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#73 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci3#10 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#74 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 devlinks.xa_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &xa->xa_lock#19 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 pin_fs_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &(&fn_net->fib_chain)->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#74 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#73 (work_completion)(&peer->transmit_handshake_work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#73 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#73 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 cb_lock genl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#33 rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &rq->__lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#6 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#15 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#16 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#16 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#16 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 &hdev->lock &____s->seqcount#2 irq_context: 0 &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#74 irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#74 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#37 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#37 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#32 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1963 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1963 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1963 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1963 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1959 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1954 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#506 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#506 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#504 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#504 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1960 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1960 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1955 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1955 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1955 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#507 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#507 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1677 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#505 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1677 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1677 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#505 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1677 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1676 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1676 irq_context: 0 kn->active#14 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#14 &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#142 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#142 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#143 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#143 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#144 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#144 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#144 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#144 &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1965 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1965 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#34 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#142 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#34 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1961 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1961 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1961 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1956 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1678 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1678 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1677 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1677 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1677 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1677 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#362 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#360 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#360 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#38 &tn->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#34 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#34 &devlink_port->type_lock irq_context: 0 rtnl_mutex hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#143 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#119 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#119 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#119 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#119 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#119 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#118 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#81 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#81 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#81 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#81 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#81 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#81 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#81 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#81 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2004 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2004 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2000 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2000 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2000 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1995 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1719 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1719 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#1719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#1719 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#1717 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#1716 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#532 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#532 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#532 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#532 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#530 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#527 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#527 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#364 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#364 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#364 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#364 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#362 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#362 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#39 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#39 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#39 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#39 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#39 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#37 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#528 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#528 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#528 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#531 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#531 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#533 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#533 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#534 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#534 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#531 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2008 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2008 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#2008 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2008 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2004 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2004 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2004 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1999 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#2009 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#2005 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2000 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#2000 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#38 rtnl_mutex &base->lock &obj_hash[i].lock